<?xml version="1.0" encoding="UTF-8"?><!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.0 20040830//EN" "journalpublishing.dtd"><article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" dtd-version="2.0" xml:lang="en" article-type="research-article"><front><journal-meta><journal-id journal-id-type="nlm-ta">JMIR Form Res</journal-id><journal-id journal-id-type="publisher-id">formative</journal-id><journal-id journal-id-type="index">27</journal-id><journal-title>JMIR Formative Research</journal-title><abbrev-journal-title>JMIR Form Res</abbrev-journal-title><issn pub-type="epub">2561-326X</issn><publisher><publisher-name>JMIR Publications</publisher-name><publisher-loc>Toronto, Canada</publisher-loc></publisher></journal-meta><article-meta><article-id pub-id-type="publisher-id">v9i1e66931</article-id><article-id pub-id-type="doi">10.2196/66931</article-id><article-categories><subj-group subj-group-type="heading"><subject>Original Paper</subject></subj-group></article-categories><title-group><article-title>Impact of AI on Breast Cancer Detection Rates in Mammography by Radiologists of Varying Experience Levels in Singapore: Preliminary Comparative Study</article-title></title-group><contrib-group><contrib contrib-type="author" equal-contrib="yes"><name name-style="western"><surname>Goh</surname><given-names>Serene Si Ning</given-names></name><degrees>MMed</degrees><xref ref-type="aff" rid="aff1">1</xref><xref ref-type="aff" rid="aff2">2</xref><xref ref-type="aff" rid="aff3">3</xref><xref ref-type="fn" rid="equal-contrib1">*</xref></contrib><contrib contrib-type="author" equal-contrib="yes"><name name-style="western"><surname>Du</surname><given-names>Hao</given-names></name><degrees>PhD</degrees><xref ref-type="aff" rid="aff1">1</xref><xref ref-type="fn" rid="equal-contrib1">*</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Tan</surname><given-names>Loon Ying</given-names></name><degrees>MMed</degrees><xref ref-type="aff" rid="aff4">4</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Seah</surname><given-names>Edward Zhen Yu</given-names></name><degrees>MMed</degrees><xref ref-type="aff" rid="aff4">4</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Lau</surname><given-names>Wai Keat</given-names></name><degrees>MMed</degrees><xref ref-type="aff" rid="aff4">4</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Ng</surname><given-names>Alvin Hong Zhi</given-names></name><degrees>MMed</degrees><xref ref-type="aff" rid="aff4">4</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Lim</surname><given-names>Shi Wei Desmond</given-names></name><degrees>MMed</degrees><xref ref-type="aff" rid="aff4">4</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Ong</surname><given-names>Han Yang</given-names></name><degrees>MMed</degrees><xref ref-type="aff" rid="aff4">4</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Lau</surname><given-names>Samuel</given-names></name><degrees>MMed</degrees><xref ref-type="aff" rid="aff4">4</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Tan</surname><given-names>Yi Liang</given-names></name><degrees>MMed</degrees><xref ref-type="aff" rid="aff4">4</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Khaw</surname><given-names>Mun sze</given-names></name><degrees>MMed</degrees><xref ref-type="aff" rid="aff4">4</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Yap</surname><given-names>Chee Woei</given-names></name><degrees>MMed</degrees><xref ref-type="aff" rid="aff4">4</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Hui</surname><given-names>Kei Yiu Douglas</given-names></name><degrees>MMed</degrees><xref ref-type="aff" rid="aff4">4</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Tan</surname><given-names>Wei Chuan</given-names></name><degrees>MMed</degrees><xref ref-type="aff" rid="aff4">4</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Abdul</surname><given-names>Haziz Siti Rozana Binti</given-names></name><degrees>MMed</degrees><xref ref-type="aff" rid="aff4">4</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Khoo</surname><given-names>Vanessa Mei Hui</given-names></name><degrees>MMed</degrees><xref ref-type="aff" rid="aff4">4</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Ge</surname><given-names>Shuliang</given-names></name><degrees>MMed</degrees><xref ref-type="aff" rid="aff4">4</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Pool</surname><given-names>Felicity Jane</given-names></name><degrees>MBChB</degrees><xref ref-type="aff" rid="aff4">4</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Choo</surname><given-names>Yun Song</given-names></name><degrees>MMed</degrees><xref ref-type="aff" rid="aff4">4</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Wang</surname><given-names>Yi</given-names></name><degrees>PhD</degrees><xref ref-type="aff" rid="aff4">4</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Jagmohan</surname><given-names>Pooja</given-names></name><degrees>MD</degrees><xref ref-type="aff" rid="aff3">3</xref><xref ref-type="aff" rid="aff4">4</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Gopinathan</surname><given-names>Premilla Pillay</given-names></name><degrees>MMed</degrees><xref ref-type="aff" rid="aff4">4</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Hartman</surname><given-names>Mikael</given-names></name><degrees>PhD</degrees><xref ref-type="aff" rid="aff1">1</xref><xref ref-type="aff" rid="aff2">2</xref><xref ref-type="aff" rid="aff3">3</xref></contrib><contrib contrib-type="author" corresp="yes"><name name-style="western"><surname>Feng</surname><given-names>Mengling</given-names></name><degrees>PhD</degrees><xref ref-type="aff" rid="aff1">1</xref><xref ref-type="aff" rid="aff5">5</xref></contrib></contrib-group><aff id="aff1"><institution>Saw Swee Hock School of Public Health, National University Health System, National University Hospital Singapore</institution><addr-line>12 Science Drive 2</addr-line><addr-line>Singapore</addr-line><country>Singapore</country></aff><aff id="aff2"><institution>Department of Breast and General Surgery, National University Health System</institution><addr-line>Singapore</addr-line><country>Singapore</country></aff><aff id="aff3"><institution>Yong Yoo Lin School of Medicine, National University of Singapore</institution><addr-line>Singapore</addr-line><country>Singapore</country></aff><aff id="aff4"><institution>Department of Diagnostic Imaging, National University Hospital</institution><addr-line>Singapore</addr-line><country>Singapore</country></aff><aff id="aff5"><institution>Institute of Data Science, National University of Singapore</institution><addr-line>Singapore</addr-line><country>Singapore</country></aff><contrib-group><contrib contrib-type="editor"><name name-style="western"><surname>Mavragani</surname><given-names>Amaryllis</given-names></name></contrib></contrib-group><contrib-group><contrib contrib-type="reviewer"><name name-style="western"><surname>Ng</surname><given-names>Qin</given-names></name></contrib><contrib contrib-type="reviewer"><name name-style="western"><surname>Pertuz</surname><given-names>Said</given-names></name></contrib></contrib-group><author-notes><corresp>Correspondence to Mengling Feng, PhD, Saw Swee Hock School of Public Health, National University Health System, National University Hospital Singapore, 12 Science Drive 2, Singapore, 117549, Singapore, 65 6516 4984; <email>ephfm@nus.edu.sg</email></corresp><fn fn-type="equal" id="equal-contrib1"><label>*</label><p>these authors contributed equally</p></fn></author-notes><pub-date pub-type="collection"><year>2025</year></pub-date><pub-date pub-type="epub"><day>24</day><month>11</month><year>2025</year></pub-date><volume>9</volume><elocation-id>e66931</elocation-id><history><date date-type="received"><day>26</day><month>09</month><year>2024</year></date><date date-type="rev-recd"><day>15</day><month>05</month><year>2025</year></date><date date-type="accepted"><day>16</day><month>05</month><year>2025</year></date></history><copyright-statement>&#x00A9; Serene Si Ning Goh, Hao Du, Loon Ying Tan, Edward Zhen Yu Seah, Wai Keat Lau, Alvin Hong Ng Ng, Shi Wei Desmond Lim, Han Yang Ong, Samuel Lau, Yi Liang Tan, Mun sze Khaw, Chee Woei Yap, Kei Yiu Douglas Hui, Wei Chuan Tan, Haziz Siti Rozana Binti Abdul, Vanessa Mei Hui Khoo, Shuliang Ge, Felicity Jane Pool, Yun Song Choo, Yi Wang, Pooja Jagmohan, Premilla Pillay Gopinathan, Mikael Hartman, Meng Ling Feng. Originally published in JMIR Formative Research (<ext-link ext-link-type="uri" xlink:href="https://formative.jmir.org">https://formative.jmir.org</ext-link>), 24.11.2025. </copyright-statement><copyright-year>2025</copyright-year><license license-type="open-access" xlink:href="https://creativecommons.org/licenses/by/4.0/"><p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (<ext-link ext-link-type="uri" xlink:href="https://creativecommons.org/licenses/by/4.0/">https://creativecommons.org/licenses/by/4.0/</ext-link>), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR Formative Research, is properly cited. The complete bibliographic information, a link to the original publication on <ext-link ext-link-type="uri" xlink:href="https://formative.jmir.org">https://formative.jmir.org</ext-link>, as well as this copyright and license information must be included.</p></license><self-uri xlink:type="simple" xlink:href="https://formative.jmir.org/2025/1/e66931"/><abstract><sec><title>Background</title><p>Breast cancer remains the most common cancer among women globally. Mammography is a key diagnostic modality; however, interpretation is increasingly challenged by rising imaging volumes, a global shortage of breast radiologists, and variability in reader experience. Artificial intelligence (AI) has been proposed as a potential adjunct to address these issues, particularly in settings with high breast density, such as Asian populations. This study aimed to evaluate the impact of AI assistance on mammographic diagnostic performance among resident and consultant radiologists in Singapore.</p></sec><sec><title>Objective</title><p>To assess whether AI assistance improves diagnostic accuracy in mammographic breast cancer detection across radiologists with varying levels of experience.</p></sec><sec sec-type="methods"><title>Methods</title><p>A multi-reader, multi-case study was conducted at the National University Hospital, Singapore, from May to August 2023. De-identified digital mammograms from 500 women (250 with cancer and 250 normal or benign) were interpreted by 17 radiologists (4 consultants, 4 senior residents, and 9 junior residents). Each radiologist read all cases over 2 reading sessions: one without AI assistance and another with AI assistance, separated by a 1-month washout period. The AI system (FxMammo) provided heatmaps and malignancy risk scores to support decision-making. Area under the curve of the receiver operating characteristic (AUROC) was used to evaluate diagnostic performance.</p></sec><sec sec-type="results"><title>Results</title><p>Among the 500 cases, 250 were malignant and 250 were non-malignant. Of the malignant cases, 16%(80/500) were ductal carcinoma in situ and 84%(420/500) were invasive cancers. Among non-malignant cases, 69.2%(346/500) were normal, 17.6%(88) benign, and 3.6%(18/500) possibly benign but stable on follow-up. Masses (54.4%, 272) and calcifications (10.8%, 54/500) were the most common findings in cancer cases. A majority of both malignant (66.8%, 334/500) and non-malignant (68%, 340/500) cases had heterogeneously or extremely dense breasts (BIRADS [Breast Imaging Reporting and Data System] categories C and D). The AI model achieved an AUROC of 0.93 (95% CI 0.91&#x2010;0.95), slightly higher than consultant radiologists (AUROC 0.90, 95% CI 0.89&#x2010;0.92; <italic>P</italic>=.21). With AI assistance, AUROC improved among junior residents (from 0.84 to 0.86; <italic>P</italic>=.38) and senior residents (from 0.85 to 0.88; <italic>P</italic>=.13), with senior residents approaching consultant-level performance (AUROC difference 0.02; <italic>P</italic>=.051). Diagnostic gains with AI were greatest in women with dense breasts and among less experienced radiologists. AI also improved inter-reader agreement and time efficiency, particularly in benign or normal cases.</p></sec><sec sec-type="conclusions"><title>Conclusions</title><p>This is the first study in Asia to evaluate AI assistance in mammography interpretation by radiologists of varying experience. AI significantly improved diagnostic performance and efficiency among residents, helping to narrow the experience-performance gap without compromising specificity. These findings suggest a role for AI in enhancing diagnostic consistency, improving workflow, and supporting training. Integration into clinical and educational settings may offer scalable benefits, though careful attention to threshold calibration, feedback loops, and real-world validation remains essential. Further studies in routine screening settings are needed to confirm generalizability and cost-effectiveness.</p></sec></abstract><kwd-group><kwd>mammography</kwd><kwd>multi-reader</kwd><kwd>multi-case</kwd><kwd>detection</kwd><kwd>breast cancer</kwd><kwd>mammograms</kwd><kwd>radiologist</kwd><kwd>radiology</kwd><kwd>Singapore</kwd><kwd>AI</kwd><kwd>artificial intelligence</kwd><kwd>cancer</kwd><kwd>women</kwd><kwd>diagnosis</kwd><kwd>diagnostic modality</kwd></kwd-group></article-meta></front><body><sec id="s1" sec-type="intro"><title>Introduction</title><p>Mammograms are a critical tool in breast cancer diagnosis; however, interpreting them is inherently challenging. Expertise is acquired only after lengthy training; however, there is a shortage of seasoned senior radiologists [<xref ref-type="bibr" rid="ref1">1</xref>] due to workforce aging and rising demand for breast cancer screening and diagnosis. The scarcity of skilled professionals in this field is particularly critical in health care systems that increasingly prioritize health screening and primary prevention. The challenge in the interpretation of mammography is further confounded when faced with dense breasts, as higher breast densities tend to obscure detection of breast lesions, a unique problem among Asian women compared to their Western counterparts [<xref ref-type="bibr" rid="ref2">2</xref>]. The introduction of artificial intelligence (AI) presents a potential solution to this issue. In recent years, there has been a growing interest in leveraging AI in medical imaging, particularly with the development and application of deep learning algorithms for digital mammography [<xref ref-type="bibr" rid="ref3">3</xref>,<xref ref-type="bibr" rid="ref4">4</xref>]. Preliminary investigations indicate that AI systems, when used as concurrent readers for mammogram interpretation, can enhance radiologist efficiency in terms of time, sensitivity, and specificity [<xref ref-type="bibr" rid="ref5">5</xref>,<xref ref-type="bibr" rid="ref6">6</xref>].</p><p>Beyond the potential workload reduction and improved cancer detection, AI features such as heatmaps and triage capabilities may provide valuable training support for resident radiologists.</p><p>Interest in AI among residents is growing, especially after its inclusion in the noninterpretive skills section of the Qualifying (Core) Exam by the American Board of Radiology in 2021 [<xref ref-type="bibr" rid="ref7">7</xref>]. The integration of AI into residency clinical workflows may extend beyond its potential role in reducing diagnostic errors; it has the potential to offer continuous mentorship, especially during times when consultants&#x2019; expertise may not be readily available. This creates a supportive and learning-oriented work environment for resident radiologists. AI could also serve as a tool for personalized precision education [<xref ref-type="bibr" rid="ref8">8</xref>], enabling each resident to accumulate expertise and receive tailored feedback to enhance their performance. As AI becomes an integral part of residency curriculums, working with it will be a crucial non-interpretive skill that residents must acquire. Looking ahead, residents will not only interpret mammographic images but also need to be attuned to unusual AI outputs, recognize and address automation bias, and understand how AI can alter clinical workflows [<xref ref-type="bibr" rid="ref9">9</xref>]. Furthermore, AI can be trained to triage benign cases, reducing the burden of routine tasks and allowing radiologists to focus on higher-value responsibilities such as analyzing complex cases. It has the potential to enhance efficiency and mitigate the risk of burnout [<xref ref-type="bibr" rid="ref10">10</xref>].</p><p>Despite the aforementioned benefits, there remains a paucity of studies evaluating the performance of AI algorithms for mammograms in the context of assisting resident radiologists. Therefore, this study is designed to investigate the performance of AI assistance for resident radiologists compared to consultant radiologists for breast cancer detection.</p></sec><sec id="s2" sec-type="methods"><title>Methods</title><sec id="s2-1"><title>Ethical Considerations</title><p>The research protocol received approval from the National Health Group Institutional Review Boards on December 16, 2022 (reference number 2022/00843). Informed consent was waived due to the use of de-identified, retrospective imaging data.</p></sec><sec id="s2-2"><title>Case Selection</title><p>A multi-reader multi-case (MRMC) investigation was conducted at the National University Hospital (NUH) from May to August 2023. The study involved a retrospective analysis of de-identified mammographic images obtained from the institutional radiology archive. All available mammographic examinations, including both screening and diagnostic mammograms conducted during the study period, were considered. Cases were identified through a systematic review of radiology reports and corresponding pathology records to ensure diagnostic confirmation and completeness. The inclusion criteria were (1) women aged 40 years and older; (2) presence of 4 standard views on mammograms; (3) availability of biopsy-proven results for malignant cases; and (4) absence of a breast cancer diagnosis after a 24-month follow-up for normal cases. Cases were excluded if they exhibited any of the following conditions: (1) evidence of mammographic needle projection, pre-biopsy markings, or clips; (2) mammographic artifacts from breast implants; (3) poor-quality mammograms; (4) malignant cases lacking corresponding biopsy-proven histology; and (5) malignant cases with histology reports exceeding 3 months from the mammogram date to ensure that cancers included were promptly detected following the mammograms. The accuracy of malignant cases was verified through biopsy results. Non-malignant cases included Breast Imaging Reporting and Data System (BIRADS) 1 normal mammogram, BIRADS 2 (benign) and BIRADS 3 (possible benign) lesions on mammograms were either biopsy-proven benign or confirmed by the absence of a breast cancer diagnosis during a 24-month follow-up period. To ensure balanced representation and maintain feasibility for a multi-reader design, the dataset was enriched to include an equal number of cancer (n=250) and non-cancer (n=250) cases, reflecting a 1:1 ratio. This approach, commonly used in diagnostic reader studies, aimed to maximize statistical power while maintaining case diversity. All images were fully anonymized prior to distribution for interpretation, and no patient-identifiable information was accessible to readers or study investigators.</p></sec><sec id="s2-3"><title>Sample Size Calculation</title><p>This retrospective, MRMC study was conducted with 500 cases and 17 radiologist readers. An equal sample of malignant cases and non-malignant cases produces for any individual reader a 2-sided 95% CI for the area under the curve of the receiver operating characteristic (AUROC) with a width (difference between lower and upper limit) no wider than 0.227. The same level of precision will be provided for specificity, as there are 250 non-malignant cases. The reader study image case set was enriched with cancer cases by having a 1:1 normal-to-malignant case ratio. This enriched case set provides an efficient and less burdensome representative case dataset. Hillis and Schartz [<xref ref-type="bibr" rid="ref11">11</xref>] recommended &#x201C;that a researcher use at least 4 readers, and preferably more, if the goal is to generalize to both reader and case populations.&#x201D; As the variability between consultant and senior resident radiologists is expected to be smaller compared to variability between resident radiologists, 4 consultants, 4 senior radiologists, and 9 resident radiologists would provide reasonably precise estimates of reader variability to prove the hypothesis.</p></sec><sec id="s2-4"><title>Reader Characteristics</title><p>In total, 17 radiologists from NUH participated in the study, comprising 4 consultants, 4 senior residents, and 9 junior resident radiologists. Resident radiologists were classified as radiology trainees who had completed the Fellowship of the Royal College of Radiologists examination and were undergoing training as part of The Accreditation Council for Graduate Medical Education radiology residency programs during the study period [<xref ref-type="bibr" rid="ref12">12</xref>,<xref ref-type="bibr" rid="ref13">13</xref>]. Senior resident radiologists were in their final year of radiology residency, while junior resident radiologists were in Years 3 and 4 of the program. Consultant radiologists were defined as those who had received accreditation from the specialty of Diagnostic Radiology of the Specialists Accreditation Board in Singapore [<xref ref-type="bibr" rid="ref14">14</xref>]. None had prior experience of reading mammograms with AI assistance in a study trial or clinical setting at the time of the study.</p></sec><sec id="s2-5"><title>AI Software</title><p>All examinations were retrospectively processed using FxMammo (version 1.1.0; FathomX) [<xref ref-type="bibr" rid="ref15">15</xref>], which has received Health Science Authority approval from Singapore. The AI system generates a continuous risk score for each view of both sides of the breast, ranging from 0% to 100%, with higher scores indicating a greater risk of suspicious findings. The highest risk score from each mammography examination was used to determine the overall patient-level risk score. The system is able to provide varying cutoff thresholds for malignancy score, which defines the levels at which the algorithm classifies an image as positive for malignant breast cancer. The reference cutoff for malignancy in this study was determined by the Youden index [<xref ref-type="bibr" rid="ref16">16</xref>]. The results are presented to the clinician as a &#x201C;heatmap,&#x201D; highlighting the regions of interest on the mammogram with their corresponding risk scores.</p></sec><sec id="s2-6"><title>Study Protocol</title><p>All radiologists read unique mammograms, acquired using the same mammographic system, of 500 women over a 6-week period without AI assistance. This was followed by a 1-month washout period and then another 6 weeks of reading the same mammograms with AI assistance. The order of the mammograms was shuffled in the second reading. All radiologists underwent a pilot trial before the study commenced, during which they were familiarized with the reading process on the OsiriX DICOM Viewer [<xref ref-type="bibr" rid="ref17">17</xref>] and the recording of their interpretations on an electronic form. The form required recording of the lesion site for suspected malignant cases (left or right breast), malignancy risk score (yes or no), and BIRADS density assessment. Readers were not informed of the cancer enrichment in the dataset. Varying cutoff thresholds of the AI, including the threshold at Youden index for malignancy score, were provided to readers so that they are able to balance their sensitivity and specificity by adjusting the criteria for classifying cases. The time taken to read each case, encompassing interpretation and form completion, was recorded by a research assistant.</p></sec><sec id="s2-7"><title>Data Analysis</title><p>Statistical analysis was performed using Python (version 3.8.16) and R (version 3.6.1, package RJafroc version 2.1.2, pROC version 1.18.4). Radiologists were grouped by their reader experience levels into junior residents, senior residents, and consultant radiologists. Sensitivity, specificity, and accuracy of malignancy risk scores were calculated in percentages. The difference in proportions between the various diagnostic metrics before and after AI assistance was compared using 2-tailed permutation tests with 2000 iterations [<xref ref-type="bibr" rid="ref18">18</xref>,<xref ref-type="bibr" rid="ref19">19</xref>]. AUROCs were calculated for each group of radiologists for malignancy risk scores assigned to each case compared to the ground truth with and without AI assistance. AUROCs were reported for each group of radiologists under each modality (eg, with or without AI assistance). For each group of radiologists, the non-parametric (trapezoidal) AUROC for the &#x201C;with AI&#x201D; read, &#x201C;without AI&#x201D; read, and the difference between them was presented. The comparison of AUROC performance in the MRMC analysis was conducted using the Obuchowski-Rockette-Hillis method [<xref ref-type="bibr" rid="ref20">20</xref>,<xref ref-type="bibr" rid="ref21">21</xref>] and DeLong&#x2019;s test [<xref ref-type="bibr" rid="ref22">22</xref>]. Covariance terms in the model were estimated using the jackknife resampling method. The jackknife method was used to estimate covariance terms in the model. Two&#x2010;sided 95% CIs were used to illustrate the precision of the within&#x2010;modality estimates and the between&#x2010;modalities difference. Inter-reader agreement in assessing cancer detection was determined using percent agreement and Cohen kappa (<italic>k</italic>). A value of <italic>P</italic>&#x003C;.05 was considered statistically significant. To account for multiple comparisons, including AUROC, accuracy, sensitivity, specificity, and Cohen kappa, the significance level was adjusted to 0.01 using the Bonferroni correction [<xref ref-type="bibr" rid="ref23">23</xref>]. Two-tailed paired t-tests were used to compare time savings before and after assistance.</p></sec><sec id="s2-8"><title>Outcomes</title><p>The primary objective of the study is to compare diagnostic accuracy in terms of cancer detection (binary outcome, yes or no cancer) between resident and consultant radiologists with and without AI assistance, measured by AUROCs. Secondary endpoints include the time taken to read each case.</p></sec></sec><sec id="s3" sec-type="results"><title>Results</title><sec id="s3-1"><title>Case Characteristics and Histology</title><p>A total of 500 women with 2000 mammographic images were eligible, including 250 non-malignant cases and 250 malignant cases (<xref ref-type="table" rid="table1">Table 1</xref>). The median age of women with malignancy was 60.2 years (IQR: 51.8-68.0) as compared to 53.0 years (IQR: 47.0 to 60.0) for women with no malignancy (<italic>P</italic>&#x003C;.001). Of the malignant cases, 16% were ductal carcinoma in situ and 84% were invasive cancer. Of the non-malignant cases, 69.2% had normal mammograms, 17.6% had benign lesions, and 3.6% had possibly benign lesions that completed 24 months of normal follow-up. Mass (54.4%) and calcification (10.8%) were the most common lesion types for malignant cases. The majority of non-malignant cases (68%) and malignant cases (66.8%) had BIRADS density categories of C and D.</p><table-wrap id="t1" position="float"><label>Table 1.</label><caption><p>Summary of clinical and histopathological characteristics of all included cases.</p></caption><table id="table1" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom">Characteristics</td><td align="left" valign="bottom">Non-malignant (n=250)</td><td align="left" valign="bottom">Malignant (n=250)</td><td align="left" valign="bottom"><italic>P</italic> value</td></tr></thead><tbody><tr><td align="left" valign="top">Age of women (median, IQR)</td><td align="left" valign="top">53.0 (47.0, 60.0)</td><td align="left" valign="top">60.2 (51.8, 68.0)</td><td align="left" valign="top">&#x003C;.001<sup><xref ref-type="table-fn" rid="table1fn1">a</xref></sup></td></tr><tr><td align="left" valign="top" colspan="4">BIRADS<sup><xref ref-type="table-fn" rid="table1fn2">b</xref></sup> density category, n (%)</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>A</td><td align="left" valign="top">8 (3.2)</td><td align="left" valign="top">7 (2.8)</td><td align="left" valign="top" rowspan="4">.496</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>B</td><td align="left" valign="top">72 (28.8)</td><td align="left" valign="top">76 (30.4)</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>C</td><td align="left" valign="top">151 (60.4)</td><td align="left" valign="top">156 (62.4)</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>D</td><td align="left" valign="top">19 (7.6)</td><td align="left" valign="top">11 (4.4)</td></tr><tr><td align="left" valign="top" colspan="4">Lesion type category, n (%)</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Mass</td><td align="left" valign="top">15 (10.1)</td><td align="left" valign="top">136 (54.4)</td><td align="left" valign="top" rowspan="6">.001<sup><xref ref-type="table-fn" rid="table1fn1">a</xref></sup></td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Calcification</td><td align="left" valign="top">12 (8.1)</td><td align="left" valign="top">27 (10.8)</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Asymmetry</td><td align="left" valign="top">98 (65.8)</td><td align="left" valign="top">21 (8.4)</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Distortion</td><td align="left" valign="top">6 (4.0)</td><td align="left" valign="top">7 (2.8)</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Mass+ calcification</td><td align="left" valign="top">2 (1.3)</td><td align="left" valign="top">27 (10.8)</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Combined lesions</td><td align="left" valign="top">16 (10.7)</td><td align="left" valign="top">32 (12.8)</td></tr><tr><td align="left" valign="top" colspan="4">BIRADS malignancy category and histopathology type category, n (%)</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Normal mammograms (BIRADS 1)</td><td align="left" valign="top">173 (69.2)</td><td align="left" valign="top">1 (0.4)</td><td align="left" valign="top">.001<sup><xref ref-type="table-fn" rid="table1fn1">a</xref></sup></td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Benign<break/>(BIRADS 2)</td><td align="left" valign="top">44 (17.6)</td><td align="left" valign="top">0 (0)</td><td align="left" valign="top">NA</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Possibly benign (BIRADS 3)</td><td align="left" valign="top">9 (3.6)</td><td align="left" valign="top">1 (0.4)</td><td align="left" valign="top">.001<sup><xref ref-type="table-fn" rid="table1fn1">a</xref></sup></td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Suspicious<break/>(BIRADS 4)</td><td align="left" valign="top">(Biopsy proven benign and concordant)<break/>24 (9.6)</td><td align="left" valign="top">160 (64.0)</td><td align="left" valign="top">.001<sup><xref ref-type="table-fn" rid="table1fn1">a</xref></sup></td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Highly suggestive of malignancy<break/>(BIRADS 5)</td><td align="left" valign="top">0 (0)</td><td align="left" valign="top">88 (35.2)</td><td align="left" valign="top">NA</td></tr></tbody></table><table-wrap-foot><fn id="table1fn1"><p><sup>a</sup>Statistically significant with <italic>P</italic> value of &#x003C;.05.</p></fn><fn id="table1fn2"><p><sup>b</sup>BIRADS: Breast Imaging Reporting and Data System.</p></fn></table-wrap-foot></table-wrap></sec><sec id="s3-2"><title>Comparison of Cancer Diagnostic Performance by Reader Experience Before and After AI Assistance</title><p>Reading with AI assistance improved the sensitivity of junior resident radiologists from 56.9% to 61.6% (difference 4.7%, 95% CI difference [1.9, 7.1% ], <italic>P</italic>&#x003C;.001) and senior resident radiologists from 55.4% to 64.1% (difference 8.7 %, 95% CI difference (4.5%, 13%) <italic>P</italic>&#x003C;.001) (<xref ref-type="fig" rid="figure1">Figure 1</xref>). Reading with AI assistance improved the sensitivity of consultant radiologists from 68.5% to 70.5% (difference 2%, 95% CI difference (&#x2212;1.9%, 6.4%), <italic>P</italic>=.35), respectively. Similarly, with AI assistance, an improved specificity was observed among junior resident radiologists (94.6%-96.3%; (difference 1.7%, 95% CI difference (0.4%, 2.9%), <italic>P</italic>=.02), senior resident radiologists (96.7%-96.7%; difference 0%, 95% CI difference (&#x2212;1.6%, 1.5%), <italic>P</italic>=.99), and consultant radiologists (96%-97%; difference 1%, 95% CI difference (&#x2212;0.6%, 2.6%), <italic>P</italic>=.22) as compared with reading without AI assistance. Junior resident radiologists had improved accuracy from 75.8% to 78.9% (difference 3.1%, 95% CI of difference (1.3%, 4.9%), <italic>P</italic>=.005) and senior resident radiologists had improved accuracy from 76.1% to 80.4% (difference 4.3%, 95% CI of difference (1.8%, 6.9%), <italic>P</italic>=.002) Consultant radiologists had improved accuracy from 82.3% to 83.9% (difference 1.6%, 95% CI of difference (0.8%, 3.8%) , <italic>P</italic>=.24).</p><fig position="float" id="figure1"><label>Figure 1.</label><caption><p>Sensitivity, specificity, and accuracy grouped by reader experience without and with artificial intelligence (AI) assistance.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="formative_v9i1e66931_fig01.png"/></fig></sec><sec id="s3-3"><title>Comparison of AUROC by Reader Experience Before and After AI Assistance</title><p>Use of the AI system standalone demonstrated an AUROC of 0.93 (95% CI 0.91-0.95), as compared to 0.90 (95% CI 0.89&#x2010;0.92) for consultant radiologists, 0.85 (95% CI 0.83-0.87) for senior resident radiologists, and 0.84 (95% CI (0.83-0.85) for junior resident radiologists.</p><p>The diagnostic performance of AI standalone is significantly higher than that of consultant radiologists (difference in AUROC 0.02, 95% CI of difference &#x2013;0.01, 0.05, <italic>P</italic>=.21) (<xref ref-type="fig" rid="figure2">Figure 2</xref>). AI&#x2019;s diagnostic performance was higher than both senior (difference in AUROC 0.06, 95% CI of difference 0.02, 0.11 <italic>P</italic>=.013) and junior residents&#x2019; (difference in AUROC 0.07, 95% CI of difference 0.05, 0.10<italic>P</italic>&#x003C;.001) assessments, respectively. There is no statistically significant difference between senior and junior residents&#x2019; performance (difference in AUROC 0.006, 95% CI of difference &#x2212;0.013, 0.026, <italic>P</italic>=.52).</p><p>With AI assistance, both junior and senior residents showed improvement in diagnostic performance, AUROC from 0.85 to 0.86 (difference 0.008, 95% CI of difference &#x2013;0.011, 0.027, <italic>P</italic>=.38), and AUROC 0.87 to 0.89 (difference 0.027, 95% CI of difference &#x2013;0.014, 0.068, <italic>P</italic>=.13), respectively. With AI assistance, the AUROC of senior residents was comparable to consultant radiologists with a difference in AUROC of 0.02 (95% CI 0.00&#x2010;0.039, <italic>P</italic>=.051). However, the diagnostic performance of junior residents remained lower than consultant radiologists despite AI assistance (difference in AUROC 0.045, 95% CI of difference 0.028, 0.061, <italic>P</italic>&#x003C;.001).</p><p>The AUROCs were lower for all reader groups when interpreting mammograms of women with high breast density compared to those with non-dense breasts, at 0.82 (95% CI 0.81&#x2010;0.83) and 0.93 (95% CI 0.92&#x2010;0.93), respectively. However, this difference was not statistically significant (<italic>P</italic>=.20). With AI assistance, the AUROC for women with dense breasts improved to 0.84 (95% CI 0.83&#x2010;0.85) from 0.82 (95% CI 0.81&#x2010;0.83), although the increase was not statistically significant (<italic>P</italic>=.13). Similar improvements were observed across all other reader groups, but none reached statistical significance (<xref ref-type="table" rid="table2">Table 2</xref>).</p><fig position="float" id="figure2"><label>Figure 2.</label><caption><p>Comparison of area under the curve of the receiver operating characteristic (AUROC) for junior and senior resident radiologists with and without artificial intelligence (AI) assistance.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="formative_v9i1e66931_fig02.png"/></fig><table-wrap id="t2" position="float"><label>Table 2.</label><caption><p>Comparison of AUROC<sup><xref ref-type="table-fn" rid="table2fn1">a</xref></sup> for reader groups based on breast density with and without AI<sup><xref ref-type="table-fn" rid="table2fn2">b</xref></sup> assistance.</p></caption><table id="table2" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom" rowspan="2">Breast density</td><td align="left" valign="bottom" colspan="4">AUROC of reader groups without AI assistance (95% CI)</td><td align="left" valign="bottom" colspan="4">AUROC of reader groups with AI assistance (95% CI)</td></tr><tr><td align="left" valign="bottom">All</td><td align="left" valign="bottom">Consultants</td><td align="left" valign="bottom">Senior residents</td><td align="left" valign="bottom">Junior residents</td><td align="left" valign="bottom">All</td><td align="left" valign="bottom">Consultants</td><td align="left" valign="bottom">Senior residents</td><td align="left" valign="bottom">Junior residents</td></tr></thead><tbody><tr><td align="left" valign="top">Non-dense breast group (density A and B)</td><td align="left" valign="top">0.93<break/>(0.92&#x2010;0.93)</td><td align="left" valign="top">0.96<break/>(0.94&#x2010;0.97)</td><td align="left" valign="top">0.93 (0.91-<break/>0.94)</td><td align="left" valign="top">0.91<break/>(0.90-<break/>0.93)</td><td align="left" valign="top">0.95<break/>(0.94&#x2010;0.95) (<italic>P</italic>=.13)<sup><xref ref-type="table-fn" rid="table2fn3">c</xref></sup></td><td align="left" valign="top">0.97<break/>(0.96-0.98) (<italic>P</italic>=.49)</td><td align="left" valign="top">0.95<break/>(0.93&#x2010;0.96) (<italic>P</italic>=.25)</td><td align="left" valign="top">0.94<break/>(0.92&#x2010;0.95) (<italic>P</italic>=.28)</td></tr><tr><td align="left" valign="top">Dense breast group (density C and D)</td><td align="left" valign="top">0.82<break/>(0.81&#x2010;0.83)</td><td align="left" valign="top">0.88<break/>(0.86&#x2010;0.90)</td><td align="left" valign="top">0.81<break/>(0.79&#x2010;0.83)</td><td align="left" valign="top">0.81<break/>(0.79-<break/>0.82)</td><td align="left" valign="top">0.84<break/>(0.83&#x2010;0.85) (<italic>P</italic>=.20)</td><td align="left" valign="top">0.88<break/>(0.86&#x2010;0.90) (<italic>P</italic>=.90)</td><td align="left" valign="top">0.85<break/>(0.87&#x2010;0.93) (<italic>P</italic>=.22)</td><td align="left" valign="top">0.82<break/>(0.81&#x2010;0.83) (<italic>P</italic>=.59)</td></tr></tbody></table><table-wrap-foot><fn id="table2fn1"><p><sup>a</sup>AUROC: area under the curve of the receiver operating characteristic.</p></fn><fn id="table2fn2"><p><sup>b</sup>AI: artificial intelligence.</p></fn><fn id="table2fn3"><p><sup>c</sup>Represents <italic>P</italic> value of &#x003C;.01 for paired groups by reader experience before and after AI assistance.</p></fn></table-wrap-foot></table-wrap></sec><sec id="s3-4"><title>Agreement Rate Between Radiologists and AI</title><p>The overall agreement rate between consultants, senior radiologists, and junior radiologists with AI was 84.1%, 81.3%, and 81.4%, respectively (<xref ref-type="table" rid="table3">Table 3</xref>). The agreement rate among junior residents before and after AI assistance was &#x03BA;=0.54 to &#x03BA;=0.60, <italic>P</italic>=.08. The agreement rate among senior residents before and after AI assistance was &#x03BA;=0.59 to &#x03BA;=0.62 <italic>P</italic>=.48. The agreement rate among consultants before and after AI assistance was 0.66 to 0.71, <italic>P</italic>=.07.</p><table-wrap id="t3" position="float"><label>Table 3.</label><caption><p>Agreement rate between AI<sup><xref ref-type="table-fn" rid="table3fn1">a</xref></sup> and radiologists.</p></caption><table id="table3" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom"/><td align="left" valign="bottom">Consultants</td><td align="left" valign="bottom">Senior residents</td><td align="left" valign="bottom">Junior residents</td></tr></thead><tbody><tr><td align="left" valign="top">Agreement rate between AI and radiologists (% proportion)</td><td align="left" valign="top">84.1</td><td align="left" valign="top">81.0</td><td align="left" valign="top">81.4</td></tr><tr><td align="left" valign="top">Discordance</td><td align="left" valign="top"/><td align="left" valign="top"/><td align="left" valign="top"/></tr><tr><td align="left" valign="top">Total number of readings across scenarios, (number of patients)</td><td align="left" valign="top">319 (145)</td><td align="left" valign="top">380 (176)</td><td align="left" valign="top">838 (220)</td></tr><tr><td align="left" valign="top">Scenario 1</td><td align="left" valign="top"/><td align="left" valign="top"/><td align="left" valign="top"/></tr><tr><td align="left" valign="top">&#x2003;AI is correct, reader false negative, number of cases (% proportion of discordant cases)</td><td align="left" valign="top">157 (49.2)</td><td align="left" valign="top">215 (56.6)</td><td align="left" valign="top">502 (59.9)</td></tr><tr><td align="left" valign="top">Scenario 2</td><td align="left" valign="top"/><td align="left" valign="top"/><td align="left" valign="top"/></tr><tr><td align="left" valign="top">&#x2003;AI is correct, reader false positive, number of cases (% proportion of discordant cases)</td><td align="left" valign="top">17 (5.3)</td><td align="left" valign="top">23 (6.1)</td><td align="left" valign="top">58 (6.9)</td></tr><tr><td align="left" valign="top">Scenario 3</td><td align="left" valign="top"/><td align="left" valign="top"/><td align="left" valign="top"/></tr><tr><td align="left" valign="top">&#x2003;Reader correct, AI false negative, number of cases (% proportion of discordant cases)</td><td align="left" valign="top">66 (20.7)</td><td align="left" valign="top">60 (15.8)</td><td align="left" valign="top">97 (11.6)</td></tr><tr><td align="left" valign="top">Scenario 4</td><td align="left" valign="top"/><td align="left" valign="top"/><td align="left" valign="top"/></tr><tr><td align="left" valign="top">&#x2003;Reader correct, AI false positive, number of cases (% proportion of discordant cases)</td><td align="left" valign="top">79 (24.8)</td><td align="left" valign="top">82 (21.6)</td><td align="left" valign="top">181 (21.6)</td></tr></tbody></table><table-wrap-foot><fn id="table3fn1"><p><sup>a</sup>AI: artificial intelligence.</p></fn></table-wrap-foot></table-wrap></sec><sec id="s3-5"><title>Reading Time Analysis</title><p>For all radiologist groups, significant time savings were achieved for both non-malignant and malignant cases following AI assistance, with the highest time savings observed in non-malignant cases (18.0 [SD 34.3] s (95% CI [17.0, 19.0]) vs malignant cases (11.1 [SD 34.2] s (95% CI [10.1, 12.1] per mammogram read after AI assistance, <italic>P</italic>&#x003C;.001). Time savings were highest for the junior residents, followed by the consultants and the senior residents (<xref ref-type="table" rid="table4">Table 4</xref>).</p><table-wrap id="t4" position="float"><label>Table 4.</label><caption><p>Time savings per mammogram read for radiologist groups with and without AI assistance.</p></caption><table id="table4" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom"/><td align="left" valign="bottom" colspan="3">Avg.<sup><xref ref-type="table-fn" rid="table4fn1">a</xref></sup> time per mammogram read without AI in seconds, mean (SD)<sup><xref ref-type="table-fn" rid="table4fn2">b</xref></sup></td><td align="left" valign="bottom" colspan="3">Avg. time per mammogram read with AI in seconds, mean (SD)</td><td align="left" valign="bottom" colspan="3">Avg. time savings per mammogram read with AI in seconds, mean (SD)</td></tr><tr><td align="left" valign="top"/><td align="left" valign="top">Consultant</td><td align="left" valign="top">Senior Resident</td><td align="left" valign="top">Junior resident</td><td align="left" valign="top">Consultant</td><td align="left" valign="top">Senior Resident</td><td align="left" valign="top">Junior resident</td><td align="left" valign="top">Consultant<break/>Time savings, 95% CI of difference</td><td align="left" valign="top">Senior Resident<break/>Time savings, 95% CI of difference</td><td align="left" valign="top">Junior resident<break/>Time savings, 95% CI of difference</td></tr></thead><tbody><tr><td align="left" valign="top">All cases</td><td align="left" valign="top">81.0<break/>(36.2)</td><td align="left" valign="top">58.2 (27.2)</td><td align="left" valign="top">70.0<break/>(41.9)</td><td align="left" valign="top">70.2<break/>(31.5)</td><td align="left" valign="top">50.1 (22.8)</td><td align="left" valign="top">50.1<break/>(29.1)</td><td align="left" valign="top">10.4 (34.0)<break/>(95% CI 8.6&#x2010;11.6), <italic>P</italic>&#x003C;.001</td><td align="left" valign="top">8.1 (27.5)<break/>(95% CI 6.9&#x2010;9.3), <italic>P</italic>&#x003C;.001</td><td align="left" valign="top">19.7 (36.8)<break/>(95% CI 18.3&#x2010;20.5), <italic>P</italic>&#x003C;.001</td></tr><tr><td align="left" valign="top">Non-malignant cases</td><td align="left" valign="top">73.1<break/>(35.1)</td><td align="left" valign="top">55.5 (26.3)</td><td align="left" valign="top">66.7 (42.4)</td><td align="left" valign="top">57.5 (26.2)</td><td align="left" valign="top">44.6 (18.9)</td><td align="left" valign="top">43.4 (28.2)</td><td align="left" valign="top">15.8 (31.8)<break/>(95% CI 12.9&#x2010;17.0), <italic>P</italic>&#x003C;0.001</td><td align="left" valign="top">11.1 (25.3)<break/>(95% CI 9.0&#x2010;12.2), <italic>P</italic>&#x003C;.001</td><td align="left" valign="top">23.4 (37.7)<break/>(95% CI 21.1&#x2010;24.2), <italic>P</italic>&#x003C;.001</td></tr><tr><td align="left" valign="top">Malignant cases</td><td align="left" valign="top">88.1<break/>(35.8)</td><td align="left" valign="top">60.0 (27.7)</td><td align="left" valign="top">73.0 (41.2)</td><td align="left" valign="top">83.0 (31.2)</td><td align="left" valign="top">55.7<break/>(25.0)</td><td align="left" valign="top">56.7 (28.4)</td><td align="left" valign="top">4.9<break/>(33.9) (95% CI 3.1&#x2010;7.3), <italic>P</italic>&#x003C;.001</td><td align="left" valign="top">5.5 (26.7) (95% CI 3.8&#x2010;7.3), <italic>P</italic>&#x003C;.001</td><td align="left" valign="top">15.4 (36.6)<break/>(95% CI 14.7&#x2010;17.7), <italic>P</italic>&#x003C;.001</td></tr></tbody></table><table-wrap-foot><fn id="table4fn1"><p><sup>a</sup>Avg.: average.</p></fn><fn id="table4fn2"><p><sup>b</sup>SD: standard deviation.</p></fn></table-wrap-foot></table-wrap></sec></sec><sec id="s4" sec-type="discussion"><title>Discussion</title><sec id="s4-1"><title>Principal Findings and Comparison With Previous Works</title><p>This study marks the first exploration of AI assistance for resident radiologists in Asia. The tested AI software exhibited a high overall AUROC performance of 0.93, aligning with results of prior research in western populations by Aggarwal et al [<xref ref-type="bibr" rid="ref24">24</xref>] and Hickman et al [<xref ref-type="bibr" rid="ref25">25</xref>], achieving AUROCs of 0.87 and 0.89, respectively. Our study demonstrated that AI assistance improved diagnostic performance across all radiologist groups, with the greatest relative gains observed among residents. While the AUROC for senior residents using AI approached that of consultants (0.88 vs 0.90, <italic>P</italic>=.051), this trend suggests the potential of AI to support less experienced readers and help narrow the performance gap between intermediate and expert radiologists, although the difference did not reach statistical significance. At least two-thirds of women in our study had high breast density. Dense breasts are known to increase the difficulty of mammographic interpretation, which is a particular challenge in Asian populations, where breast density tends to be higher compared to Western counterparts [<xref ref-type="bibr" rid="ref2">2</xref>,<xref ref-type="bibr" rid="ref26">26</xref>]. With AI assistance, the cancer detection rate on mammograms of women with high breast density showed improvement across all radiologist groups, with senior residents experiencing the highest improvement, reflected in an AUROC increase from 81% to 85%. The ability of any AI software to deliver diagnostic performance comparable to that of a second reader in double-reading systems introduces the potential to streamline workflows, reducing the reliance on a second reader and alleviating radiologists&#x2019; workload. However, to fully realize this potential, rigorous testing of various thresholds and continuous evaluations with feedback loops will be necessary. These measures will ensure the system evolves over time and remains aligned with clinical outcomes, maintaining its utility and effectiveness in real-world settings [<xref ref-type="bibr" rid="ref27">27</xref>].</p><p>Furthermore, AI assistance was shown to improve sensitivity and accuracy for radiologists at all levels, particularly benefiting junior and senior residents. This supports the idea that AI can help mitigate the effects of lower diagnostic experience, which is commonly seen in radiology residents who have not yet accrued a high volume of cases. The study by Agarwal et al [<xref ref-type="bibr" rid="ref28">28</xref>] underscores the correlation between a resident&#x2019;s diagnostic performance and the number of studies they interpret during their training, making the assistance of AI a valuable asset in overcoming this learning curve. For consultant radiologists, the gains in accuracy were less pronounced, suggesting that the advantage of AI assistance is more substantial for those earlier in their careers. This phenomenon might be explained by the &#x201C;regression to the mean&#x201D; effect, where extreme performances (eg, by junior radiologists with less experience) tend to improve with intervention (AI assistance), while those with already higher baseline performance (experienced consultants) see more modest improvements [<xref ref-type="bibr" rid="ref29">29</xref>]. While AI enhances overall diagnostic capabilities, its transformative impact is most evident at lower experience levels, where there is more room for improvement.</p><p>Beyond individual benefits, incorporating AI into structured residency curricula focused on practical applications offers an opportunity to enhance diagnostic proficiency, accelerate skill acquisition, and foster confidence in early-career clinicians. This preparation is essential for adapting to the rapidly evolving and digitizing field of medical imaging [<xref ref-type="bibr" rid="ref30">30</xref>]. A nationwide cross-sectional survey by Chen et al, involving 3666 participants, revealed that radiology residents generally hold positive attitudes toward AI, with 72.8% acknowledging its potential to improve disease diagnosis and 78.18% emphasizing the importance of embracing AI in practice. However, 29.9% expressed concerns about AI reducing the demand for radiologists. Maria et al [<xref ref-type="bibr" rid="ref31">31</xref>] described the feasibility of a 3-day AI curriculum that successfully improved radiologists&#x2019; perceptions of their knowledge and skills regarding AI in radiology, serving as a potential model for further adaptation and implementation in educational settings.</p><p>The observation that both junior and senior radiologists demonstrated lower sensitivity but higher specificity compared to the AI software underscores key differences in diagnostic strategies between human radiologists and AI systems. With AI assistance, however, both groups showed a significant improvement in sensitivity without sacrificing specificity. This suggests that radiologists may naturally prioritize specificity to avoid false positives and unnecessary interventions, which explains their higher specificity. However, this cautious approach can lead to reduced sensitivity, potentially resulting in missed diagnoses (false negatives), particularly in subtle or complex cases. In contrast, AI systems are typically designed to maintain high sensitivity thresholds, leveraging their capability to identify subtle abnormalities that might be overlooked by human readers. This distinction highlights the complementary strengths of human radiologists and AI [<xref ref-type="bibr" rid="ref32">32</xref>]. Hence, integrating AI into clinical workflows may help achieve a better balance, enhancing radiologists&#x2019; sensitivity without compromising specificity, ultimately improving diagnostic accuracy. Enhanced sensitivity through AI integration can help reduce false negatives, leading to earlier detection and treatment of conditions like breast cancer. At the same time, maintaining high specificity reduces unnecessary biopsies, alleviating patient discomfort and lowering health care costs [<xref ref-type="bibr" rid="ref33">33</xref>].</p><p>Next, AI assistance was found to improve efficiency, particularly through time savings, with longer savings observed in non-malignant cases. This indicates that the AI software could offer substantial cost savings by enhancing efficiency, especially in screening settings where the ratio of non-cancer to cancer cases is typically higher [<xref ref-type="bibr" rid="ref34">34</xref>]. Consistent with cost-effectiveness analyses of AI solutions in other domains, the greatest time savings were seen in non-specialist groups [<xref ref-type="bibr" rid="ref35">35</xref>] rather than expert radiologists, where the marginal advantages are limited. Furthermore, AI assistance improved inter-reader agreement across all radiologist groups, with junior residents benefiting the most. This highlights AI&#x2019;s potential to reduce variability in diagnostic decisions and enhance consistency among radiologists of varying experience levels. However, the lack of perfect agreement between radiologists and AI indicates that discrepancies in interpretation persist. These disagreements often lead to additional time spent analyzing differences, further investigations, or the need for arbitration by additional personnel. Mehrizi et al [<xref ref-type="bibr" rid="ref36">36</xref>] observed that when AI provided an incorrect suggestion but radiologists ultimately made the correct decision, these tasks took longer, averaging 89 seconds compared to the overall average of 79 seconds. Therefore, the thoughtful integration of AI into clinical workflows is important to ensure that its benefits, such as improved efficiency and consistency, outweigh any added complexities or challenges.</p><p>The strength of this study lies in its provision of valuable insights into how AI can complement human decision-making and enhance diagnostic performance at different stages of professional development. Nonetheless, several limitations should be acknowledged. First, the dataset was enriched to a 50% cancer prevalence, consistent with previous multi-reader studies [<xref ref-type="bibr" rid="ref3">3</xref>,<xref ref-type="bibr" rid="ref4">4</xref>], to reduce the total number of cases required and minimize the burden on the 17 participating readers while retaining the ability to assess diagnostic accuracy. However, as real-world screening populations typically have lower cancer prevalence, there is concern that the AI system may exhibit a higher false positive rate in clinical practice. In addition, the lack of clinical variables, such as patient age, prior imaging, and treatment history, may limit the model&#x2019;s robustness and reduce its generalizability across diverse patient populations. While the study was not specifically powered for formal hypothesis testing across all reader subgroups, the inclusion of 4 consultants, 4 senior residents, and 9 junior residents exceeds the minimum threshold recommended by Hillis and Schartz [<xref ref-type="bibr" rid="ref11">11</xref>] for generalizing to both reader and case populations. This permitted exploratory analyses of variability in diagnostic performance by reader experience level. Due to logistical constraints, a crossover design was not implemented, introducing the possibility of unintended training effects, where improved reading efficiency during the second round may have been due in part to increased familiarity rather than the AI intervention. To mitigate this, several strategies were used: a pilot session was conducted to familiarize readers with the AI tool; case order was shuffled between rounds; and a washout period was introduced. These measures were designed to minimize practice effects and enhance the attribution of performance changes to the AI tool itself.</p><p>Despite these limitations, this study remains a crucial initial step in assessing the performance of AI software in supporting resident radiologists, as well as in the context of dense Asian breasts. The study process revealed insights into potential challenges that could emerge during the initial phases of integrating AI [<xref ref-type="bibr" rid="ref37">37</xref>], such as the need for interpretive adjustment and the issue of misdiagnosis, which requires careful consideration. Clear definitions of the roles and accountabilities of AI, senior residents, and consultants in the event of misdiagnosis are essential [<xref ref-type="bibr" rid="ref38">38</xref>]. Future endeavors will include an in-depth analysis of the discordance cases between radiologist groups and the AI solution, understanding perspectives of human readers and establishing a feedback loop for model improvement to ensure that AI becomes a valuable tool in health care while maintaining patient safety and quality of care. Lastly, a prospective multi-center collaboration would provide a larger and more diverse dataset, representative of the Asian population, to assess the AI system&#x2019;s effectiveness across a broader range of cases and demographics [<xref ref-type="bibr" rid="ref39">39</xref>].</p><sec id="s4-1-1"><title>Conclusions</title><p>In summary, this study provides preliminary insights into the use of AI assistance in mammographic interpretation among radiologists in Singapore. While AI showed some potential in improving sensitivity, particularly for junior and senior residents, its overall impact on diagnostic accuracy was modest, with many comparisons not reaching statistical significance. Automation bias and inter-reader variability remain challenges that require careful consideration. The enriched dataset and absence of a crossover design are also some of the limitations of our study. Future studies should prioritize multi-center collaborations, external validation, and in-depth analyses of discordant cases to ensure AI integration enhances diagnostic workflows while maintaining safety and clinical relevance.</p></sec></sec></sec></body><back><ack><p>We wish to express our gratitude to our dedicated research and administrative staff&#x2014;Jenny Liu, Nur Khaliesah Binte Mohamed Riza, Ganga Devi D/O Chandrasegran, and Mei Ying Ng&#x2014;for their contribution to the preparation and execution of the project. This study was supported by Temasek Foundation and Ministry of Health Singapore through the CHI START UP ENTERPRISE LINK (CHISEL) program Program 2.0 (NCSP 2.0) under the Department of Surgery, National University Hospital, and Breast Cancer Screening Prevention Programme (NUHSRO/2020/121/BCSPP/LOA) under the Yong Loo Lin School of Medicine, National University of Singapore. The grant was used to fund the manpower required for the project. The study design, data collection, data analysis, and interpretation were conducted by the study team independently.</p></ack><notes><sec><title>Data Availability</title><p>The datasets generated and analyzed during the present study are not publicly available due to institutional data governance policies and patient privacy regulations. The de-identified mammographic images and associated diagnostic data are restricted to approved research use under the ethics protocol (DSRB reference no. 2022/00843). Reasonable requests for access to anonymized data for academic or collaborative purposes may be directed to the corresponding author and will require approval from the relevant institutional review board.</p></sec></notes><fn-group><fn fn-type="conflict"><p>MLF and MH are co-founders of FathomX. However, this research is not directly related to the commercial activities of FathomX. The study was conducted with transparency and impartiality, in collaboration with the NUH study group.</p></fn></fn-group><glossary><title>Abbreviations</title><def-list><def-item><term id="abb1">AI</term><def><p>artificial intelligence</p></def></def-item><def-item><term id="abb2">AUROC</term><def><p>area under curve of the receiver operating characteristic</p></def></def-item><def-item><term id="abb3">BIRADS</term><def><p>Breast Imaging Reporting and Data System</p></def></def-item><def-item><term id="abb4">MRMC</term><def><p>multi-reader multi-case</p></def></def-item><def-item><term id="abb5">NUH</term><def><p>National University Hospital</p></def></def-item></def-list></glossary><ref-list><title>References</title><ref id="ref1"><label>1</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Lai</surname><given-names>AYT</given-names> </name></person-group><article-title>The growing problem of radiologist shortage: Hong Kong&#x2019;s perspective</article-title><source>Korean J Radiol</source><year>2023</year><month>10</month><volume>24</volume><issue>10</issue><fpage>931</fpage><lpage>932</lpage><pub-id pub-id-type="doi">10.3348/kjr.2023.0838</pub-id><pub-id pub-id-type="medline">37793661</pub-id></nlm-citation></ref><ref id="ref2"><label>2</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Maskarinec</surname><given-names>G</given-names> </name><name name-style="western"><surname>Meng</surname><given-names>L</given-names> </name><name name-style="western"><surname>Ursin</surname><given-names>G</given-names> </name></person-group><article-title>Ethnic differences in mammographic densities</article-title><source>Int J Epidemiol</source><year>2001</year><month>10</month><volume>30</volume><issue>5</issue><fpage>959</fpage><lpage>965</lpage><pub-id pub-id-type="doi">10.1093/ije/30.5.959</pub-id><pub-id pub-id-type="medline">11689504</pub-id></nlm-citation></ref><ref id="ref3"><label>3</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Rodr&#x00ED;guez-Ruiz</surname><given-names>A</given-names> </name><name name-style="western"><surname>Krupinski</surname><given-names>E</given-names> </name><name name-style="western"><surname>Mordang</surname><given-names>JJ</given-names> </name><etal/></person-group><article-title>Detection of breast cancer with mammography: effect of an artificial intelligence support system</article-title><source>Radiology</source><year>2019</year><month>02</month><volume>290</volume><issue>2</issue><fpage>305</fpage><lpage>314</lpage><pub-id pub-id-type="doi">10.1148/radiol.2018181371</pub-id><pub-id pub-id-type="medline">30457482</pub-id></nlm-citation></ref><ref id="ref4"><label>4</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Pacil&#x00E8;</surname><given-names>S</given-names> </name><name name-style="western"><surname>Lopez</surname><given-names>J</given-names> </name><name name-style="western"><surname>Chone</surname><given-names>P</given-names> </name><name name-style="western"><surname>Bertinotti</surname><given-names>T</given-names> </name><name name-style="western"><surname>Grouin</surname><given-names>JM</given-names> </name><name name-style="western"><surname>Fillard</surname><given-names>P</given-names> </name></person-group><article-title>Improving breast cancer detection accuracy of mammography with the concurrent use of an artificial intelligence tool</article-title><source>Radiol Artif Intell</source><year>2020</year><month>11</month><volume>2</volume><issue>6</issue><fpage>e190208</fpage><pub-id pub-id-type="doi">10.1148/ryai.2020190208</pub-id><pub-id pub-id-type="medline">33937844</pub-id></nlm-citation></ref><ref id="ref5"><label>5</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Raya-Povedano</surname><given-names>JL</given-names> </name><name name-style="western"><surname>Romero-Mart&#x00ED;n</surname><given-names>S</given-names> </name><name name-style="western"><surname>El&#x00ED;as-Cabot</surname><given-names>E</given-names> </name><name name-style="western"><surname>Gubern-M&#x00E9;rida</surname><given-names>A</given-names> </name><name name-style="western"><surname>Rodr&#x00ED;guez-Ruiz</surname><given-names>A</given-names> </name><name name-style="western"><surname>&#x00C1;lvarez-Benito</surname><given-names>M</given-names> </name></person-group><article-title>AI-based strategies to reduce workload in breast cancer screening with mammography and tomosynthesis: a retrospective evaluation</article-title><source>Radiology</source><year>2021</year><month>07</month><volume>300</volume><issue>1</issue><fpage>57</fpage><lpage>65</lpage><pub-id pub-id-type="doi">10.1148/radiol.2021203555</pub-id><pub-id pub-id-type="medline">33944627</pub-id></nlm-citation></ref><ref id="ref6"><label>6</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>L&#x00E5;ng</surname><given-names>K</given-names> </name><name name-style="western"><surname>Dustler</surname><given-names>M</given-names> </name><name name-style="western"><surname>Dahlblom</surname><given-names>V</given-names> </name><name name-style="western"><surname>&#x00C5;kesson</surname><given-names>A</given-names> </name><name name-style="western"><surname>Andersson</surname><given-names>I</given-names> </name><name name-style="western"><surname>Zackrisson</surname><given-names>S</given-names> </name></person-group><article-title>Identifying normal mammograms in a large screening population using artificial intelligence</article-title><source>Eur Radiol</source><year>2021</year><month>03</month><volume>31</volume><issue>3</issue><fpage>1687</fpage><lpage>1692</lpage><pub-id pub-id-type="doi">10.1007/s00330-020-07165-1</pub-id><pub-id pub-id-type="medline">32876835</pub-id></nlm-citation></ref><ref id="ref7"><label>7</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Garin</surname><given-names>SP</given-names> </name><name name-style="western"><surname>Zhang</surname><given-names>V</given-names> </name><name name-style="western"><surname>Jeudy</surname><given-names>J</given-names> </name><name name-style="western"><surname>Parekh</surname><given-names>VS</given-names> </name><name name-style="western"><surname>Yi</surname><given-names>PH</given-names> </name></person-group><article-title>Systematic review of radiology residency artificial intelligence curricula: preparing future radiologists for the artificial intelligence era</article-title><source>J Am Coll Radiol</source><year>2023</year><month>06</month><volume>20</volume><issue>6</issue><fpage>561</fpage><lpage>569</lpage><pub-id pub-id-type="doi">10.1016/j.jacr.2023.02.031</pub-id><pub-id pub-id-type="medline">37127217</pub-id></nlm-citation></ref><ref id="ref8"><label>8</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Wang</surname><given-names>C</given-names> </name><name name-style="western"><surname>Xie</surname><given-names>H</given-names> </name><name name-style="western"><surname>Wang</surname><given-names>S</given-names> </name><name name-style="western"><surname>Yang</surname><given-names>S</given-names> </name><name name-style="western"><surname>Hu</surname><given-names>L</given-names> </name></person-group><article-title>Radiological education in the era of artificial intelligence: a review</article-title><source>Medicine (Baltimore)</source><year>2023</year><month>01</month><day>6</day><volume>102</volume><issue>1</issue><fpage>e32518</fpage><pub-id pub-id-type="doi">10.1097/MD.0000000000032518</pub-id><pub-id pub-id-type="medline">36607870</pub-id></nlm-citation></ref><ref id="ref9"><label>9</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Dratsch</surname><given-names>T</given-names> </name><name name-style="western"><surname>Chen</surname><given-names>X</given-names> </name><name name-style="western"><surname>Rezazade Mehrizi</surname><given-names>M</given-names> </name><etal/></person-group><article-title>Automation bias in mammography: the impact of artificial intelligence BI-RADS suggestions on reader performance</article-title><source>Radiology</source><year>2023</year><month>05</month><volume>307</volume><issue>4</issue><fpage>e222176</fpage><pub-id pub-id-type="doi">10.1148/radiol.222176</pub-id><pub-id pub-id-type="medline">37129490</pub-id></nlm-citation></ref><ref id="ref10"><label>10</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Kalantarova</surname><given-names>S</given-names> </name><name name-style="western"><surname>Mickinac</surname><given-names>N</given-names> </name><name name-style="western"><surname>Santhosh</surname><given-names>S</given-names> </name><etal/></person-group><article-title>Preventing physician burnout in breast imaging: scope of the problem and keys to success</article-title><source>Curr Probl Diagn Radiol</source><year>2021</year><volume>50</volume><issue>5</issue><fpage>734</fpage><lpage>737</lpage><pub-id pub-id-type="doi">10.1067/j.cpradiol.2020.09.005</pub-id><pub-id pub-id-type="medline">33069520</pub-id></nlm-citation></ref><ref id="ref11"><label>11</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Hillis</surname><given-names>SL</given-names> </name><name name-style="western"><surname>Schartz</surname><given-names>KM</given-names> </name></person-group><article-title>Demonstration of multi- and single-reader sample size program for diagnostic studies software</article-title><source>Proc SPIE Int Soc Opt Eng</source><year>2015</year><month>02</month><volume>9416</volume><fpage>94160E</fpage><pub-id pub-id-type="doi">10.1117/12.2083150</pub-id><pub-id pub-id-type="medline">28890592</pub-id></nlm-citation></ref><ref id="ref12"><label>12</label><nlm-citation citation-type="web"><article-title>The Royal College of Radiologists</article-title><source>Joint Final Clinical Radiology exams Joint Final Clinical Radiology Exams</source><year>2023</year><access-date>2023-10-10</access-date><publisher-name>The Royal College of Radiologists</publisher-name><comment><ext-link ext-link-type="uri" xlink:href="https://www.rcr.ac.uk/clinical-radiology/exams/joint-final-clinical-radiology-exams">https://www.rcr.ac.uk/clinical-radiology/exams/joint-final-clinical-radiology-exams</ext-link></comment></nlm-citation></ref><ref id="ref13"><label>13</label><nlm-citation citation-type="web"><person-group person-group-type="author"><collab>Accreditation Council for Graduate Medical Education (ACGME)</collab></person-group><article-title>ACGME program requirements for graduate medical education in diagnostic radiology</article-title><year>2023</year><access-date>2023-10-10</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://www.acgme.org/globalassets/pfassets/programrequirements/2025-reformatted-requirements/420_diagnosticradiology_2025_reformatted.pdf">https://www.acgme.org/globalassets/pfassets/programrequirements/2025-reformatted-requirements/420_diagnosticradiology_2025_reformatted.pdf</ext-link></comment></nlm-citation></ref><ref id="ref14"><label>14</label><nlm-citation citation-type="web"><person-group person-group-type="author"><collab>Specialists Accreditation Board (SAB)</collab></person-group><source>Diagnostic Radiology Residency</source><year>2022</year><month>09</month><day>1</day><access-date>2023-10-10</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://sab.healthprofessionals.gov.sg/specialist-training/list-of-recognised-specialties/diagnostic-radiology/">https://sab.healthprofessionals.gov.sg/specialist-training/list-of-recognised-specialties/diagnostic-radiology/</ext-link></comment></nlm-citation></ref><ref id="ref15"><label>15</label><nlm-citation citation-type="web"><article-title>FathomX</article-title><source>Home</source><year>2023</year><access-date>2023-10-10</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://www.fathomx.co/">https://www.fathomx.co/</ext-link></comment></nlm-citation></ref><ref id="ref16"><label>16</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Ruopp</surname><given-names>MD</given-names> </name><name name-style="western"><surname>Perkins</surname><given-names>NJ</given-names> </name><name name-style="western"><surname>Whitcomb</surname><given-names>BW</given-names> </name><name name-style="western"><surname>Schisterman</surname><given-names>EF</given-names> </name></person-group><article-title>Youden Index and optimal cut-point estimated from observations affected by a lower limit of detection</article-title><source>Biom J</source><year>2008</year><month>06</month><volume>50</volume><issue>3</issue><fpage>419</fpage><lpage>430</lpage><pub-id pub-id-type="doi">10.1002/bimj.200710415</pub-id><pub-id pub-id-type="medline">18435502</pub-id></nlm-citation></ref><ref id="ref17"><label>17</label><nlm-citation citation-type="web"><person-group person-group-type="author"><collab>OsiriX</collab></person-group><article-title>OsiriX DICOM viewer</article-title><source>The world famous medical imaging viewer</source><year>2023</year><access-date>2023-10-10</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://www.osirix-viewer.com/">https://www.osirix-viewer.com/</ext-link></comment></nlm-citation></ref><ref id="ref18"><label>18</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Nakai</surname><given-names>E</given-names> </name><name name-style="western"><surname>Miyagi</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Suzuki</surname><given-names>K</given-names> </name><etal/></person-group><article-title>Artificial intelligence as a second reader for screening mammography</article-title><source>Radiol Adv</source><year>2024</year><month>05</month><day>28</day><volume>1</volume><issue>2</issue><fpage>umae011</fpage><pub-id pub-id-type="doi">10.1093/radadv/umae011</pub-id></nlm-citation></ref><ref id="ref19"><label>19</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>McKinney</surname><given-names>SM</given-names> </name><name name-style="western"><surname>Sieniek</surname><given-names>M</given-names> </name><name name-style="western"><surname>Godbole</surname><given-names>V</given-names> </name><etal/></person-group><article-title>International evaluation of an AI system for breast cancer screening</article-title><source>Nature New Biol</source><year>2020</year><month>01</month><volume>577</volume><issue>7788</issue><fpage>89</fpage><lpage>94</lpage><pub-id pub-id-type="doi">10.1038/s41586-019-1799-6</pub-id><pub-id pub-id-type="medline">31894144</pub-id></nlm-citation></ref><ref id="ref20"><label>20</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Obuchowski</surname><given-names>NA</given-names> </name><name name-style="western"><surname>Rockette</surname><given-names>HE</given-names> </name></person-group><article-title>Hypothesis testing of diagnostic accuracy for multiple readers and multiple tests an anova approach with dependent observations</article-title><source>Commun Statist Simul Comput</source><year>1995</year><month>01</month><volume>24</volume><issue>2</issue><fpage>285</fpage><lpage>308</lpage><pub-id pub-id-type="doi">10.1080/03610919508813243</pub-id></nlm-citation></ref><ref id="ref21"><label>21</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Hillis</surname><given-names>SL</given-names> </name></person-group><article-title>A comparison of denominator degrees of freedom methods for multiple observer ROC analysis</article-title><source>Stat Med</source><year>2007</year><month>02</month><day>10</day><volume>26</volume><issue>3</issue><fpage>596</fpage><lpage>619</lpage><pub-id pub-id-type="doi">10.1002/sim.2532</pub-id><pub-id pub-id-type="medline">16538699</pub-id></nlm-citation></ref><ref id="ref22"><label>22</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>DeLong</surname><given-names>ER</given-names> </name><name name-style="western"><surname>DeLong</surname><given-names>DM</given-names> </name><name name-style="western"><surname>Clarke-Pearson</surname><given-names>DL</given-names> </name></person-group><article-title>Comparing the areas under two or more correlated receiver operating characteristic curves: a nonparametric approach</article-title><source>Biometrics</source><year>1988</year><month>09</month><volume>44</volume><issue>3</issue><fpage>837</fpage><lpage>845</lpage><pub-id pub-id-type="medline">3203132</pub-id></nlm-citation></ref><ref id="ref23"><label>23</label><nlm-citation citation-type="book"><person-group person-group-type="author"><name name-style="western"><surname>Rupert Jr</surname><given-names>G</given-names> </name></person-group><source>Simultaneous Statistical Inference</source><year>2012</year></nlm-citation></ref><ref id="ref24"><label>24</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Aggarwal</surname><given-names>R</given-names> </name><name name-style="western"><surname>Sounderajah</surname><given-names>V</given-names> </name><name name-style="western"><surname>Martin</surname><given-names>G</given-names> </name><etal/></person-group><article-title>Diagnostic accuracy of deep learning in medical imaging: a systematic review and meta-analysis</article-title><source>NPJ Digit Med</source><year>2021</year><month>04</month><day>7</day><volume>4</volume><issue>1</issue><fpage>65</fpage><pub-id pub-id-type="doi">10.1038/s41746-021-00438-z</pub-id><pub-id pub-id-type="medline">33828217</pub-id></nlm-citation></ref><ref id="ref25"><label>25</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Hickman</surname><given-names>SE</given-names> </name><name name-style="western"><surname>Woitek</surname><given-names>R</given-names> </name><name name-style="western"><surname>Le</surname><given-names>EPV</given-names> </name><etal/></person-group><article-title>Machine learning for workflow applications in screening mammography: systematic review and meta-analysis</article-title><source>Radiology</source><year>2022</year><month>01</month><volume>302</volume><issue>1</issue><fpage>88</fpage><lpage>104</lpage><pub-id pub-id-type="doi">10.1148/radiol.2021210391</pub-id><pub-id pub-id-type="medline">34665034</pub-id></nlm-citation></ref><ref id="ref26"><label>26</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Lee</surname><given-names>SE</given-names> </name><name name-style="western"><surname>Yoon</surname><given-names>JH</given-names> </name><name name-style="western"><surname>Son</surname><given-names>NH</given-names> </name><name name-style="western"><surname>Han</surname><given-names>K</given-names> </name><name name-style="western"><surname>Moon</surname><given-names>HJ</given-names> </name></person-group><article-title>Screening in patients with dense breasts: comparison of mammography, artificial intelligence, and supplementary ultrasound</article-title><source>AJR Am J Roentgenol</source><year>2024</year><month>01</month><volume>222</volume><issue>1</issue><fpage>1</fpage><lpage>11</lpage><pub-id pub-id-type="doi">10.2214/AJR.23.29655</pub-id><pub-id pub-id-type="medline">37493324</pub-id></nlm-citation></ref><ref id="ref27"><label>27</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Goh</surname><given-names>SSN</given-names> </name><name name-style="western"><surname>Hartman</surname><given-names>M</given-names> </name></person-group><article-title>From promise to practice: harnessing artificial intelligence for breast cancer screening</article-title><source>Intell Oncol</source><year>2024</year><month>11</month><day>22</day><comment><ext-link ext-link-type="uri" xlink:href="https://www.sciencedirect.com/science/article/pii/S2950261624000025">https://www.sciencedirect.com/science/article/pii/S2950261624000025</ext-link></comment></nlm-citation></ref><ref id="ref28"><label>28</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Agarwal</surname><given-names>V</given-names> </name><name name-style="western"><surname>Bump</surname><given-names>GM</given-names> </name><name name-style="western"><surname>Heller</surname><given-names>MT</given-names> </name><etal/></person-group><article-title>Resident case volume correlates with clinical performance: finding the sweet spot</article-title><source>Acad Radiol</source><year>2019</year><month>01</month><volume>26</volume><issue>1</issue><fpage>136</fpage><lpage>140</lpage><pub-id pub-id-type="doi">10.1016/j.acra.2018.06.023</pub-id><pub-id pub-id-type="medline">30087064</pub-id></nlm-citation></ref><ref id="ref29"><label>29</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Yu</surname><given-names>F</given-names> </name><name name-style="western"><surname>Moehring</surname><given-names>A</given-names> </name><name name-style="western"><surname>Banerjee</surname><given-names>O</given-names> </name><name name-style="western"><surname>Salz</surname><given-names>T</given-names> </name><name name-style="western"><surname>Agarwal</surname><given-names>N</given-names> </name><name name-style="western"><surname>Rajpurkar</surname><given-names>P</given-names> </name></person-group><article-title>Heterogeneity and predictors of the effects of AI assistance on radiologists</article-title><source>Nat Med</source><year>2024</year><month>03</month><access-date>2024-03-20</access-date><volume>30</volume><issue>3</issue><fpage>837</fpage><lpage>849</lpage><comment><ext-link ext-link-type="uri" xlink:href="https://www.nature.com/articles/s41591-024-02850-w">https://www.nature.com/articles/s41591-024-02850-w</ext-link></comment><pub-id pub-id-type="doi">10.1038/s41591-024-02850-w</pub-id><pub-id pub-id-type="medline">38504016</pub-id></nlm-citation></ref><ref id="ref30"><label>30</label><nlm-citation citation-type="other"><person-group person-group-type="author"><name name-style="western"><surname>Chen</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Wu</surname><given-names>Z</given-names> </name><name name-style="western"><surname>Wang</surname><given-names>P</given-names> </name><name name-style="western"><surname>Xie</surname><given-names>L</given-names> </name><name name-style="western"><surname>Yan</surname><given-names>M</given-names> </name><name name-style="western"><surname>Jiang</surname><given-names>M</given-names> </name><etal/></person-group><article-title>Ready to embrace artificial intelligence? results of a nationwide survey on radiology residents in China: a cross-sectional study</article-title><source>J Med Internet Res</source><comment>Preprint posted online on  Oct 19, 2023</comment><pub-id pub-id-type="doi">10.2196/48249</pub-id></nlm-citation></ref><ref id="ref31"><label>31</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>van Kooten</surname><given-names>MJ</given-names> </name><name name-style="western"><surname>Tan</surname><given-names>CO</given-names> </name><name name-style="western"><surname>Hofmeijer</surname><given-names>EIS</given-names> </name><etal/></person-group><article-title>A framework to integrate artificial intelligence training into radiology residency programs: preparing the future radiologist</article-title><source>Insights Imaging</source><year>2024</year><month>01</month><day>17</day><volume>15</volume><issue>1</issue><pub-id pub-id-type="doi">10.1186/s13244-023-01595-3</pub-id></nlm-citation></ref><ref id="ref32"><label>32</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Leibig</surname><given-names>C</given-names> </name><name name-style="western"><surname>Brehmer</surname><given-names>M</given-names> </name><name name-style="western"><surname>Bunk</surname><given-names>S</given-names> </name><name name-style="western"><surname>Byng</surname><given-names>D</given-names> </name><name name-style="western"><surname>Pinker</surname><given-names>K</given-names> </name><name name-style="western"><surname>Umutlu</surname><given-names>L</given-names> </name></person-group><article-title>Combining the strengths of radiologists and AI for breast cancer screening: a retrospective analysis</article-title><source>Lancet Digit Health</source><year>2022</year><month>07</month><volume>4</volume><issue>7</issue><fpage>e507</fpage><lpage>e519</lpage><pub-id pub-id-type="doi">10.1016/S2589-7500(22)00070-X</pub-id><pub-id pub-id-type="medline">35750400</pub-id></nlm-citation></ref><ref id="ref33"><label>33</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Zeng</surname><given-names>A</given-names> </name><name name-style="western"><surname>Houssami</surname><given-names>N</given-names> </name><name name-style="western"><surname>Noguchi</surname><given-names>N</given-names> </name><name name-style="western"><surname>Nickel</surname><given-names>B</given-names> </name><name name-style="western"><surname>Marinovich</surname><given-names>ML</given-names> </name></person-group><article-title>Frequency and characteristics of errors by artificial intelligence (AI) in reading screening mammography: a systematic review</article-title><source>Breast Cancer Res Treat</source><year>2024</year><month>08</month><volume>207</volume><issue>1</issue><fpage>1</fpage><lpage>13</lpage><pub-id pub-id-type="doi">10.1007/s10549-024-07353-3</pub-id><pub-id pub-id-type="medline">38853221</pub-id></nlm-citation></ref><ref id="ref34"><label>34</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Tajmir</surname><given-names>SH</given-names> </name><name name-style="western"><surname>Alkasab</surname><given-names>TK</given-names> </name></person-group><article-title>Toward augmented radiologists: changes in radiology education in the era of machine learning and artificial intelligence</article-title><source>Acad Radiol</source><year>2018</year><month>06</month><volume>25</volume><issue>6</issue><fpage>747</fpage><lpage>750</lpage><pub-id pub-id-type="doi">10.1016/j.acra.2018.03.007</pub-id><pub-id pub-id-type="medline">29599010</pub-id></nlm-citation></ref><ref id="ref35"><label>35</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Vargas-Palacios</surname><given-names>A</given-names> </name><name name-style="western"><surname>Sharma</surname><given-names>N</given-names> </name><name name-style="western"><surname>Sagoo</surname><given-names>GS</given-names> </name></person-group><article-title>Cost-effectiveness requirements for implementing artificial intelligence technology in the Women&#x2019;s UK Breast Cancer Screening service</article-title><source>Nat Commun</source><year>2023</year><month>09</month><day>30</day><volume>14</volume><issue>1</issue><fpage>6110</fpage><pub-id pub-id-type="doi">10.1038/s41467-023-41754-0</pub-id><pub-id pub-id-type="medline">37777510</pub-id></nlm-citation></ref><ref id="ref36"><label>36</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Rezazade Mehrizi</surname><given-names>MH</given-names> </name><name name-style="western"><surname>Mol</surname><given-names>F</given-names> </name><name name-style="western"><surname>Peter</surname><given-names>M</given-names> </name><etal/></person-group><article-title>The impact of AI suggestions on radiologists&#x2019; decisions: a pilot study of explainability and attitudinal priming interventions in mammography examination</article-title><source>Sci Rep</source><year>2023</year><month>06</month><day>7</day><access-date>2023-06-30</access-date><volume>13</volume><issue>1</issue><fpage>9230</fpage><comment><ext-link ext-link-type="uri" xlink:href="https://www.nature.com/articles/s41598-023-36435-3">https://www.nature.com/articles/s41598-023-36435-3</ext-link></comment><pub-id pub-id-type="doi">10.1038/s41598-023-36435-3</pub-id><pub-id pub-id-type="medline">37286665</pub-id></nlm-citation></ref><ref id="ref37"><label>37</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Bahl</surname><given-names>M</given-names> </name></person-group><article-title>Artificial intelligence: a primer for breast imaging radiologists</article-title><source>J Breast Imaging</source><year>2020</year><month>08</month><volume>2</volume><issue>4</issue><fpage>304</fpage><lpage>314</lpage><pub-id pub-id-type="doi">10.1093/jbi/wbaa033</pub-id><pub-id pub-id-type="medline">32803154</pub-id></nlm-citation></ref><ref id="ref38"><label>38</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Lokaj</surname><given-names>B</given-names> </name><name name-style="western"><surname>Pugliese</surname><given-names>MT</given-names> </name><name name-style="western"><surname>Kinkel</surname><given-names>K</given-names> </name><name name-style="western"><surname>Lovis</surname><given-names>C</given-names> </name><name name-style="western"><surname>Schmid</surname><given-names>J</given-names> </name></person-group><article-title>Barriers and facilitators of artificial intelligence conception and implementation for breast imaging diagnosis in clinical practice: a scoping review</article-title><source>Eur Radiol</source><year>2024</year><month>03</month><volume>34</volume><issue>3</issue><fpage>2096</fpage><lpage>2109</lpage><pub-id pub-id-type="doi">10.1007/s00330-023-10181-6</pub-id><pub-id pub-id-type="medline">37658895</pub-id></nlm-citation></ref><ref id="ref39"><label>39</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Kelly</surname><given-names>CJ</given-names> </name><name name-style="western"><surname>Karthikesalingam</surname><given-names>A</given-names> </name><name name-style="western"><surname>Suleyman</surname><given-names>M</given-names> </name><name name-style="western"><surname>Corrado</surname><given-names>G</given-names> </name><name name-style="western"><surname>King</surname><given-names>D</given-names> </name></person-group><article-title>Key challenges for delivering clinical impact with artificial intelligence</article-title><source>BMC Med</source><year>2019</year><month>10</month><day>29</day><volume>17</volume><issue>1</issue><fpage>195</fpage><pub-id pub-id-type="doi">10.1186/s12916-019-1426-2</pub-id><pub-id pub-id-type="medline">31665002</pub-id></nlm-citation></ref></ref-list><app-group><supplementary-material id="app1"><label>Checklist 1</label><p>STARD checklist.</p><media xlink:href="formative_v9i1e66931_app1.docx" xlink:title="DOCX File, 20 KB"/></supplementary-material></app-group></back></article>