<?xml version="1.0" encoding="UTF-8"?><!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.0 20040830//EN" "journalpublishing.dtd"><article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" dtd-version="2.0" xml:lang="en" article-type="research-article"><front><journal-meta><journal-id journal-id-type="nlm-ta">JMIR Form Res</journal-id><journal-id journal-id-type="publisher-id">formative</journal-id><journal-id journal-id-type="index">27</journal-id><journal-title>JMIR Formative Research</journal-title><abbrev-journal-title>JMIR Form Res</abbrev-journal-title><issn pub-type="epub">2561-326X</issn><publisher><publisher-name>JMIR Publications</publisher-name><publisher-loc>Toronto, Canada</publisher-loc></publisher></journal-meta><article-meta><article-id pub-id-type="publisher-id">v9i1e76097</article-id><article-id pub-id-type="doi">10.2196/76097</article-id><article-categories><subj-group subj-group-type="heading"><subject>Original Paper</subject></subj-group></article-categories><title-group><article-title>Evaluating the Quality and Understandability of Radiology Report Summaries Generated by ChatGPT: Survey Study</article-title></title-group><contrib-group><contrib contrib-type="author"><name name-style="western"><surname>Sunshine</surname><given-names>Alexis</given-names></name><degrees>MD</degrees><xref ref-type="aff" rid="aff1">1</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Honce</surname><given-names>Grace H</given-names></name><degrees>MPH</degrees><xref ref-type="aff" rid="aff2">2</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Callen</surname><given-names>Andrew L</given-names></name><degrees>MD</degrees><xref ref-type="aff" rid="aff1">1</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Zander</surname><given-names>David A</given-names></name><degrees>MD</degrees><xref ref-type="aff" rid="aff1">1</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Tanabe</surname><given-names>Jody L</given-names></name><degrees>MD</degrees><xref ref-type="aff" rid="aff1">1</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Pisani Petrucci</surname><given-names>Samantha L</given-names></name><degrees>MD</degrees><xref ref-type="aff" rid="aff1">1</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Lin</surname><given-names>Chen-Tan</given-names></name><degrees>MD</degrees><xref ref-type="aff" rid="aff3">3</xref></contrib><contrib contrib-type="author" corresp="yes"><name name-style="western"><surname>Honce</surname><given-names>Justin M</given-names></name><degrees>MD</degrees><xref ref-type="aff" rid="aff1">1</xref></contrib></contrib-group><aff id="aff1"><institution>Department of Radiology, University of Colorado Anschutz Medical Campus</institution><addr-line>19th Ave. Mail Stop C278</addr-line><addr-line>Aurora</addr-line><addr-line>CO</addr-line><country>United States</country></aff><aff id="aff2"><institution>Hartway Evaluation Group</institution><addr-line>Denver</addr-line><addr-line>CO</addr-line><country>United States</country></aff><aff id="aff3"><institution>Department of Medicine, University of Colorado Anschutz Medical Campus</institution><addr-line>Aurora</addr-line><addr-line>CO</addr-line><country>United States</country></aff><contrib-group><contrib contrib-type="editor"><name name-style="western"><surname>Mavragani</surname><given-names>Amaryllis</given-names></name></contrib></contrib-group><contrib-group><contrib contrib-type="reviewer"><name name-style="western"><surname>Emekli</surname><given-names>Emre</given-names></name></contrib><contrib contrib-type="reviewer"><name name-style="western"><surname>Hershenhouse</surname><given-names>Jacob</given-names></name></contrib></contrib-group><author-notes><corresp>Correspondence to Justin M Honce, MD, Department of Radiology, University of Colorado Anschutz Medical Campus, 19th Ave. Mail Stop C278, Aurora, CO, 80045, United States, 1 303-724-3796, 1 303-724-3795; <email>justin.honce@cuanschutz.edu</email></corresp></author-notes><pub-date pub-type="collection"><year>2025</year></pub-date><pub-date pub-type="epub"><day>27</day><month>8</month><year>2025</year></pub-date><volume>9</volume><elocation-id>e76097</elocation-id><history><date date-type="received"><day>16</day><month>04</month><year>2025</year></date><date date-type="rev-recd"><day>20</day><month>06</month><year>2025</year></date><date date-type="accepted"><day>27</day><month>06</month><year>2025</year></date></history><copyright-statement>&#x00A9; Alexis Sunshine, Grace H Honce, Andrew L Callen, David A Zander, Jody L Tanabe, Samantha L Pisani Petrucci, Chen-Tan Lin, Justin M Honce. Originally published in JMIR Formative Research (<ext-link ext-link-type="uri" xlink:href="https://formative.jmir.org">https://formative.jmir.org</ext-link>), 27.8.2025. </copyright-statement><copyright-year>2025</copyright-year><license license-type="open-access" xlink:href="https://creativecommons.org/licenses/by/4.0/"><p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (<ext-link ext-link-type="uri" xlink:href="https://creativecommons.org/licenses/by/4.0/">https://creativecommons.org/licenses/by/4.0/</ext-link>), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR Formative Research, is properly cited. The complete bibliographic information, a link to the original publication on <ext-link ext-link-type="uri" xlink:href="https://formative.jmir.org">https://formative.jmir.org</ext-link>, as well as this copyright and license information must be included.</p></license><self-uri xlink:type="simple" xlink:href="https://formative.jmir.org/2025/1/e76097"/><abstract><sec><title>Background</title><p>Radiology reports convey critical medical information to health care providers and patients. Unfortunately, they are often difficult for patients to comprehend, causing confusion and anxiety, thereby limiting patient engagement in health care decision-making. Large language models (LLMs) like ChatGPT (OpenAI) can create simplified, patient-friendly report summaries to increase accessibility, albeit with errors.</p></sec><sec><title>Objective</title><p>We evaluated the accuracy and clarity of ChatGPT-generated summaries compared to original radiologist-assessed radiology reports, assessed patients&#x2019; understanding and satisfaction with the summaries compared to the original reports, and compared the readability of the original reports and summaries using validated readability metrics.</p></sec><sec sec-type="methods"><title>Methods</title><p>We anonymized 30 radiology reports created by neuroradiologists at our institution (6 brain magnetic resonance imaging, 6 brain computed tomography, 6 head and neck computed tomography angiography, 6 neck computed tomography, and 6 spine computed tomography). These anonymized reports were processed by ChatGPT to produce patient-centric summaries. Four board-certified neuroradiologists evaluated the ChatGPT-generated summaries on quality and accuracy compared to the original reports, and 4 patient volunteers separately evaluated the reports and summaries on perceived understandability and satisfaction. Readability was assessed using word count and validated readability scales.</p></sec><sec sec-type="results"><title>Results</title><p>After reading the summary, patient confidence in understanding (98%, 116/118 vs 26%, 31/118) and satisfaction regarding the level of jargon/terminology (91%, 107/118 vs 8%, 9/118) and time taken to understand the content (97%, 115/118 vs 23%, 27/118) substantially improved. Ninety-two percent (108/118) of responses indicated the summary clarified patients&#x2019; questions about the report, and 98% (116/118) of responses indicated patients would use the summary if available, with 67% (79/118) of responses indicating they would want access to both the report and summary, while 26% (31/118) of responses indicated only wanting the summary. Eighty-three percent (100/120) of radiologist responses indicated the summary represented the original report &#x201C;extremely well&#x201D; or &#x201C;very well,&#x201D; with only 5% (6/120) of responses indicating it did so &#x201C;slightly well&#x201D; or &#x201C;not well at all.&#x201D; Five percent (6/120) of responses indicated there was missing relevant medical information in the summary, 12% (14/120) reported instances of overemphasis of nonsignificant findings, and 18% (22/120) reported instances of underemphasis of significant findings. No fabricated findings were identified. Overall, 83% (99/120) of responses indicated that the summary would definitely/probably not lead patients to incorrect conclusions about the original report, with 10% (12/120) of responses indicating the summaries may do so.</p></sec><sec sec-type="conclusions"><title>Conclusions</title><p>ChatGPT-generated summaries could significantly improve perceived comprehension and satisfaction while accurately reflecting most key information from original radiology reports. Instances of minor omissions and under-/overemphasis were noted in some summaries, underscoring the need for ongoing validation and oversight. Overall, these artificial intelligence&#x2013;generated, patient-centric summaries hold promise for enhancing patient-centered communication in radiology.</p></sec></abstract><kwd-group><kwd>large language models</kwd><kwd>artificial intelligence</kwd><kwd>ChatGPT</kwd><kwd>radiology reports</kwd><kwd>patient-centered care</kwd><kwd>neuroradiology</kwd><kwd>report summarization</kwd></kwd-group></article-meta></front><body><sec id="s1" sec-type="intro"><title>Introduction</title><p>The 21st Century Cures Act increased patients&#x2019; access to their medical information [<xref ref-type="bibr" rid="ref1">1</xref>,<xref ref-type="bibr" rid="ref2">2</xref>]. This change reflected a growing desire among patients to engage with their care, a desire that is exaggerated in the field of radiology, where patients want access to reports and images, and access these materials at higher rates than general patient portals or clinical notes [<xref ref-type="bibr" rid="ref3">3</xref>-<xref ref-type="bibr" rid="ref6">6</xref>]. Unfortunately, radiology reports, which are primarily written for referring clinicians, use complex medical terminology that is well above the established guidelines for patient-friendly material [<xref ref-type="bibr" rid="ref1">1</xref>-<xref ref-type="bibr" rid="ref4">4</xref>,<xref ref-type="bibr" rid="ref7">7</xref>-<xref ref-type="bibr" rid="ref9">9</xref>]. Limited comprehension of these reports can leave patients feeling disempowered, anxious, and uncertain about their diagnoses and treatment plans [<xref ref-type="bibr" rid="ref4">4</xref>]. This is particularly worrisome given that patient engagement with and comprehension of their health care information correlates with improved patient experience, improved outcomes, and decreased costs [<xref ref-type="bibr" rid="ref1">1</xref>,<xref ref-type="bibr" rid="ref3">3</xref>,<xref ref-type="bibr" rid="ref8">8</xref>,<xref ref-type="bibr" rid="ref10">10</xref>-<xref ref-type="bibr" rid="ref12">12</xref>]. This evolving medical landscape necessitates urgent adaptations within radiology to provide easily understandable, patient-friendly reports [<xref ref-type="bibr" rid="ref1">1</xref>,<xref ref-type="bibr" rid="ref2">2</xref>].</p><p>There have been several attempts at improving the readability of radiology reports. For example, the University of Pennsylvania&#x2019;s PORTER platform is a structured reporting system that provides automated lay-language translation of reports, and University of California, San Diego, has implemented a consumer health vocabulary automated translation and embedded hyperlinks for defining key terms [<xref ref-type="bibr" rid="ref4">4</xref>,<xref ref-type="bibr" rid="ref13">13</xref>,<xref ref-type="bibr" rid="ref14">14</xref>]. However, there is a need to find ways to simplify language and avoid jargon to facilitate clear patient communication [<xref ref-type="bibr" rid="ref4">4</xref>,<xref ref-type="bibr" rid="ref8">8</xref>,<xref ref-type="bibr" rid="ref9">9</xref>,<xref ref-type="bibr" rid="ref9">9</xref>,<xref ref-type="bibr" rid="ref15">15</xref>,<xref ref-type="bibr" rid="ref16">16</xref>]. Radiologists face increasing workloads, making the manual preparation of simplified reports for every patient impractical [<xref ref-type="bibr" rid="ref17">17</xref>]. Recent improvements in artificial intelligence (AI), particularly large language models (LLMs) such as ChatGPT (OpenAI), have emerged as a promising solution for generating accessible and context-appropriate textual summaries [<xref ref-type="bibr" rid="ref2">2</xref>]. These models show promise in efficiently processing complex medical information and producing coherent summaries that minimize technical jargon while preserving essential clinical content [<xref ref-type="bibr" rid="ref18">18</xref>,<xref ref-type="bibr" rid="ref19">19</xref>].</p><p>The ability of LLMs to generate patient-centric summaries may help address critical barriers to patient engagement, specifically low health literacy levels and limited health care professional time for conversational patient education [<xref ref-type="bibr" rid="ref8">8</xref>,<xref ref-type="bibr" rid="ref12">12</xref>]. Automating this process with an LLM offers a potential solution that could be integrated into clinical workflows while preserving radiologists&#x2019; time for critical interpretive tasks. However, important concerns persist regarding the completeness and accuracy of LLM-generated outputs, biases in the underlying training data, and the consequences of generating misleading clinical information. Therefore, effective deployment of such technology requires rigorous assessment of, and balance between, the quality and simplicity of these summaries, as well as their actual benefit to patients [<xref ref-type="bibr" rid="ref20">20</xref>,<xref ref-type="bibr" rid="ref21">21</xref>].</p><p>This study evaluates ChatGPT&#x2019;s capacity to produce patient-centered radiology report summaries through three primary aims: (1) assessing the technical accuracy and completeness of the summaries by a radiologist, (2) measuring improvements in patient comprehension and satisfaction, and (3) quantifying changes in readability using validated metrics. Through this evaluation, we aimed to provide insights into how LLM-driven text generation could advance patient-centered radiology while maintaining clinical accuracy, ultimately contributing to broader discussions about the role of AI in health care communication.</p></sec><sec id="s2" sec-type="methods"><title>Methods</title><sec id="s2-1"><title>Radiology Report Collection and Anonymization</title><p>The electronic medical record was queried for radiology reports performed at our institution in September 2023. A total of 30 radiology reports were selected for analysis (6 brain magnetic resonance imaging, 6 brain computed tomography, 6 head and neck computed tomography angiography, 6 neck computed tomography, and 6 lumbar spine computed tomography) to reflect a broad range of imaging types and pathologies encountered in our neuroradiology section. Each radiology report was anonymized to remove all identifying information, including dates of comparison studies, if referenced.</p></sec><sec id="s2-2"><title>Patient-Centric Summary Creation</title><p>Each radiology report was processed by ChatGPT-4.0 (OpenAI), accessed via the web interface in October 2023, within a new chat dialogue to ensure consistent performance of the model across each report. No custom instructions or memory features were activated; thus, the prompt template for each report conversion followed a zero-shot prompt format, as follows:</p><disp-quote><p>Please rewrite the following radiology report as a patient-friendly summary between 100 and 175 words. Start with a simple overall impression statement using natural language that is easy to understand. Focus on concisely summarizing the most relevant findings, conclusions, and recommended next steps. Minimize medical jargon, but clearly define any complex technical terms needed for accuracy. Provide enough details and context so that a layperson understands the key results and implications of the radiology report. Accurately reflect any potential abnormalities rather than omitting them. The summary should strike a balance between brevity and inclusion of significant information from the original report. The goal is an accessible and easy to understand summary that avoids oversimplification and remains clinically aligned with objective findings. Ensure the significance and severity of any abnormalities are accurately represented without downplaying or exaggerating. Only include next steps if you are highly confident they are appropriate recommendations for the patient.</p></disp-quote></sec><sec id="s2-3"><title>Recruitment</title><p>Members of the study team (JMH and AS) visited our health system&#x2019;s Patient and Family Advisory Council regularly scheduled meetings to outline the purpose and structure of our study and its focus on evaluating AI-generated summaries of radiology reports, and to ask for volunteers to participate. Four patients subsequently volunteered to complete survey questions about the reports and summaries. Four board-certified neuroradiologists were directly recruited by the study team (JMH) to complete survey questions on the quality and accuracy of the summaries from a single institution.</p></sec><sec id="s2-4"><title>Survey Design</title><p>The patient survey was a closed survey designed to separately evaluate patients&#x2019; understanding, satisfaction, and perceived utility of the radiology reports and summaries. The radiologist survey was a closed survey designed to assess the technical accuracy, clinical completeness, clarity, and potential errors that may have been introduced into the summary. Both surveys included 5-point Likert-scale questions and open-ended responses. Surveys were created and administered electronically to all study participants using Qualtrics software, with each participant being sent a unique link to prevent multiple submissions. In the survey, each report and summary was clearly labeled as &#x201C;Original Radiology Report&#x201D; and &#x201C;ChatGPT Patient-Centric Summary,&#x201D; respectively. The usability of the survey instruments was tested by the study team before deployment using Qualtrics. Complete survey instruments are provided in <xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref>.</p></sec><sec id="s2-5"><title>Word Count and Readability Analysis</title><p>Textual characteristics of the original radiology reports and patient-centric summaries were analyzed via word count and validated readability metrics. These included the Flesch-Kincaid Grade Level, Flesch Reading Ease score, Gunning-Fog Index, and SMOG (Simple Measure of Gobbledygook) Index, which provide estimates of the complexity and accessibility of the reports and summaries [<xref ref-type="bibr" rid="ref22">22</xref>-<xref ref-type="bibr" rid="ref24">24</xref>].</p></sec><sec id="s2-6"><title>Statistical Analysis</title><p>Differences in word count and readability scores between the reports and summaries were calculated with paired <italic>t</italic> tests using the Microsoft Excel Analysis ToolPak. For the survey questions, closed-ended responses were rated on Likert scales, with scores reported as percentages, and reflected as top 2 (or bottom 2) box scores as appropriate, aggregating responses from the Likert scale. Each patient and radiologist&#x2019;s response for each report and summary was considered an individual data point.</p><p>Data from the open-ended questions were analyzed using a multifaceted qualitative approach. To capture a comprehensive understanding of patient and provider perspectives, the method of analysis was based on the structure of each question. During inductive analysis, open-ended responses were coded manually, sorting responses by found themes. To ensure that information was not lost or assumptions about implied importance, each response could be coded to multiple themes.</p><p>To determine how well patients understood the original radiology report and summary, their responses to the question &#x201C;What do you believe to be the most important/main point that the radiologist was trying to convey?&#x201D; were directly compared for congruence with answers supplied by the radiologist for each report/summary pair. Comparisons were then categorized as congruent or incongruent with the radiologist&#x2019;s interpretation, which was considered the gold standard.</p><p>In addition, patient responses after reading the summary were compared to their answers after reading only the report. If the answers from the summary matched the answer from the report, responses were considered congruent (ie, the summary did not change the patient&#x2019;s subjective understanding). If the answers did not match, responses were considered incongruent (ie, the summary did change the patient&#x2019;s subjective understanding), with an additional positive or negative change modifier. A positive change modifier was assigned when responses included additional details or demonstrated closer alignment with the radiologist&#x2019;s responses. Conversely, a negative change modifier was assigned when responses contained fewer details or diverged further from the radiologist&#x2019;s responses.</p></sec><sec id="s2-7"><title>Ethical Considerations</title><p>This study evaluated the quality and understandability of ChatGPT-generated radiology report summaries through both quantitative and qualitative analyses via surveys. Ethics approval was granted by the Colorado Multiple institutional review board (23&#x2010;0731). All study participants were fully informed about this study, potential risks, and their right to withdraw at any time. All participants provided electronic informed consent. No compensation was provided to participants. All data were deidentified, and all analyses were conducted in accordance with data privacy guidelines.</p></sec></sec><sec id="s3" sec-type="results"><title>Results</title><sec id="s3-1"><title>Readability</title><p>Patient-centric summaries demonstrated significant improvements in readability compared to original radiology reports across all measured metrics (<xref ref-type="table" rid="table1">Table 1</xref>). The summaries achieved a mean Flesch-Kincaid Grade Level of 7.2, representing a middle school reading level, compared to 12.5 for the original reports (<italic>P</italic>&#x003C;.001). Similar improvements were observed in Flesch Reading Ease score, Gunning-Fog Index, and SMOG Index, consistently indicating enhanced accessibility of the summary content.</p><table-wrap id="t1" position="float"><label>Table 1.</label><caption><p>Validated readability metric scores for the original radiology reports and ChatGPT-generated summaries.</p></caption><table id="table1" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom">Metric</td><td align="left" valign="bottom">Original report, mean (SD)</td><td align="left" valign="bottom">Summaries, mean (SD)</td><td align="left" valign="bottom"><italic>P</italic> value</td></tr></thead><tbody><tr><td align="left" valign="top">Word count</td><td align="left" valign="top">241.0 (130.5)</td><td align="left" valign="top">189.0 (34.7)</td><td align="left" valign="top">.02</td></tr><tr><td align="left" valign="top">Flesch-Kincaid Grade Level</td><td align="left" valign="top">12.5 (1.63)</td><td align="left" valign="top">7.2 (1.14)</td><td align="left" valign="top">&#x003C;.001</td></tr><tr><td align="left" valign="top">Flesch Reading Ease score</td><td align="left" valign="top">21.5 (11.39)</td><td align="left" valign="top">69.8 (6.2)</td><td align="left" valign="top">&#x003C;.001</td></tr><tr><td align="left" valign="top">Gunning-Fog Index</td><td align="left" valign="top">16.4 (2.38)</td><td align="left" valign="top">10.3 (1.41)</td><td align="left" valign="top">&#x003C;.001</td></tr><tr><td align="left" valign="top">SMOG<sup><xref ref-type="table-fn" rid="table1fn1">a</xref></sup> Index</td><td align="left" valign="top">12.7 (1.55)</td><td align="left" valign="top">10.4 (1.05)</td><td align="left" valign="top">&#x003C;.001</td></tr></tbody></table><table-wrap-foot><fn id="table1fn1"><p><sup>a</sup>SMOG: Simple Measure of Gobbledygook.</p></fn></table-wrap-foot></table-wrap></sec><sec id="s3-2"><title>Radiologist Survey</title><p>Four board-certified neuroradiologists completed surveys for each of the 30 radiology report/summary pairs, resulting in a total of 120 responses per question. The neuroradiologists had a mean of 11.6 (SD 9.33, range 3.5&#x2010;25) years of experience in neuroradiology after completing a fellowship.</p><p>Radiologist assessments of the summaries were generally favorable (<xref ref-type="table" rid="table2">Table 2</xref>). In 83% (100/120) of responses, radiologists indicated that the summaries represented the original reports extremely well/very well, and 90% (108/120) were satisfied with the balance of medical jargon to patient-centric language. Regarding accuracy, 5% (6/120) of responses indicated missing information, 12% (14/120) indicated overemphasis of findings, and 18% (22/120) indicated underemphasis. Notably, no hallucinations or fabricated findings were identified in any summary. Regarding the potential for patient misunderstanding, 82% (99/120) of radiologist responses indicated the summaries would likely not lead to incorrect conclusions, while 10% (12/120) indicated a perceived risk of misinterpretation.</p><table-wrap id="t2" position="float"><label>Table 2.</label><caption><p>Radiologist assessment of the quality and accuracy of ChatGPT-generated summaries (N=120).</p></caption><table id="table2" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom">Measure and response</td><td align="left" valign="bottom">Value, n (%)</td></tr></thead><tbody><tr><td align="left" valign="top" colspan="2">Representation of original findings</td></tr><tr><td align="left" valign="top">&#x2003;Extremely well/very well</td><td align="left" valign="top">100 (83)</td></tr><tr><td align="left" valign="top">&#x2003;Neutral</td><td align="left" valign="top">14 (12)</td></tr><tr><td align="left" valign="top">&#x2003;Slightly well/not well at all</td><td align="left" valign="top">6 (5)</td></tr><tr><td align="left" valign="top" colspan="2">Satisfaction with the balance of medical jargon and patient-centric language</td></tr><tr><td align="left" valign="top">&#x2003;Very satisfied/somewhat satisfied</td><td align="left" valign="top">108 (90)</td></tr><tr><td align="left" valign="top">&#x2003;Neutral</td><td align="left" valign="top">5 (4)</td></tr><tr><td align="left" valign="top">&#x2003;Somewhat dissatisfied/very dissatisfied</td><td align="left" valign="top">7 (6)</td></tr><tr><td align="left" valign="top" colspan="2">Error analysis</td></tr><tr><td align="left" valign="top">&#x2003;Missing relevant information</td><td align="left" valign="top">6 (5)</td></tr><tr><td align="left" valign="top">&#x2003;Overemphasis of findings</td><td align="left" valign="top">14 (12)</td></tr><tr><td align="left" valign="top">&#x2003;Underemphasis of findings</td><td align="left" valign="top">22 (18)</td></tr><tr><td align="left" valign="top" colspan="2">Risk of leading patients to incorrect conclusions</td></tr><tr><td align="left" valign="top">&#x2003;Definitely yes/probably yes</td><td align="left" valign="top">12 (10)</td></tr><tr><td align="left" valign="top">&#x2003;Neutral</td><td align="left" valign="top">9 (8)</td></tr><tr><td align="left" valign="top">&#x2003;Probably not/definitely not</td><td align="left" valign="top">99 (82)</td></tr></tbody></table></table-wrap><p>In their free-text responses, neuroradiologists indicated several issues with the summaries. They noted an overemphasis on minor findings (eg, mild sinus disease, white matter changes, and renal cysts) and perceived some follow-up recommendations as unnecessary. Radiologists also reported disagreement with specific terminology choices in some cases (eg, using &#x201C;mildly enlarged&#x201D; to describe a &#x201C;stable to incrementally&#x201D; enlarged collection). They further noted that summaries underemphasized some critical findings, such as arterial occlusions, perfusion deficits, and nerve root compression, by not adequately conveying their clinical significance. In addition, radiologists expressed concerns that oversimplified vocabulary diminished the importance of certain findings, such as describing a vertebral artery as &#x201C;a smaller artery in your neck&#x201D; or referring to a normalized brain volume at the 6th percentile as &#x201C;a bit below average.&#x201D;</p><p>Qualitative analysis of radiologist feedback revealed several themes regarding desired improvements to the summaries. Fifty-four percent (66/122) of free-text responses indicated no improvements were needed. In 38% (46/122), radiologists expressed discomfort with the perceived oversimplification of language (eg, &#x201C;lump&#x201D; instead of &#x201C;nodule,&#x201D; &#x201C;thinning&#x201D; instead of &#x201C;osteopenia,&#x201D; &#x201C;squishing&#x201D; instead of &#x201C;height loss&#x201D;), and generally found the summaries could have been more concise. In 4% (5/122) of responses, radiologists indicated a preference for more detail, including more clearly defined terminology and more explicit localization of relevant findings (eg, the level of spinal stenoses).</p></sec><sec id="s3-3"><title>Patient Survey</title><p>Four patient volunteers completed surveys for each of the 30 report-summary pairs, resulting in a total of 120 responses across all 30 reports and summaries for each question. A total of 118 responses were valid, with the remaining 2 responses left blank. Patients&#x2019; reported age ranged between 55 and 84 years: 1 (55-64 years), 2 (65-74 years), and 1 (75-84 years). Patients&#x2019; reported educational attainment included high school graduate (n=1), college degree (n=2), and master&#x2019;s degree (n=1).</p><p>Patient survey responses indicated substantially improved subjective comprehension and satisfaction with summaries compared to the original reports (<xref ref-type="table" rid="table3">Table 3</xref>). There were large, highly significant improvements in patient ratings when reading the summary versus the original report (<italic>P</italic>&#x003C;.001). While only 26% (31/118) of responses indicated confidence in understanding the original reports, this increased to 98% (116/118) of responses for summaries. Satisfaction with medical terminology showed similar improvement, rising from 8% (9/118) for original reports to 91% (107/118) for summaries. Time required for comprehension was a significant concern with original reports, with only 23% (27/118) satisfied with the time needed for understanding, compared to 97% (115/118) satisfaction with summaries.</p><table-wrap id="t3" position="float"><label>Table 3.</label><caption><p>Patient confidence and satisfaction with the original radiology reports and ChatGPT-generated summaries (N=118).</p></caption><table id="table3" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom">Measure and response</td><td align="left" valign="bottom">Original report, n (%)</td><td align="left" valign="bottom">Summaries, n (%)</td><td align="left" valign="bottom"><italic>P</italic> value</td></tr></thead><tbody><tr><td align="left" valign="top" colspan="3">Confidence in understanding</td><td align="left" valign="top">&#x003C;.001</td></tr><tr><td align="left" valign="top">&#x2003;Very confident/somewhat confident</td><td align="left" valign="top">31 (26)</td><td align="left" valign="top">116 (98)</td><td align="left" valign="top"/></tr><tr><td align="left" valign="top">&#x2003;Neutral</td><td align="left" valign="top">14 (12)</td><td align="left" valign="top">1 (1)</td><td align="left" valign="top"/></tr><tr><td align="left" valign="top">&#x2003;Very unconfident/somewhat unconfident</td><td align="left" valign="top">73 (62)</td><td align="left" valign="top">1 (1)</td><td align="left" valign="top"/></tr><tr><td align="left" valign="top" colspan="3">Satisfaction with medical jargon/terminology</td><td align="left" valign="top">&#x003C;.001</td></tr><tr><td align="left" valign="top">&#x2003;Very satisfied/somewhat satisfied</td><td align="left" valign="top">9 (8)</td><td align="left" valign="top">107 (91)</td><td align="left" valign="top"/></tr><tr><td align="left" valign="top">&#x2003;Neutral</td><td align="left" valign="top">18 (15)</td><td align="left" valign="top">9 (8)</td><td align="left" valign="top"/></tr><tr><td align="left" valign="top">&#x2003;Very dissatisfied/somewhat dissatisfied</td><td align="left" valign="top">91 (77)</td><td align="left" valign="top">2 (2)</td><td align="left" valign="top"/></tr><tr><td align="left" valign="top" colspan="3">Satisfaction with level of detail</td><td align="left" valign="top">&#x003C;.001</td></tr><tr><td align="left" valign="top">&#x2003;Very satisfied/somewhat satisfied</td><td align="left" valign="top">51 (43)</td><td align="left" valign="top">116 (98)</td><td align="left" valign="top"/></tr><tr><td align="left" valign="top">&#x2003;Neutral</td><td align="left" valign="top">43 (36)</td><td align="left" valign="top">1 (1)</td><td align="left" valign="top"/></tr><tr><td align="left" valign="top">&#x2003;Very dissatisfied/somewhat dissatisfied</td><td align="left" valign="top">24 (20)</td><td align="left" valign="top">1 (1)</td><td align="left" valign="top"/></tr><tr><td align="left" valign="top" colspan="3">Satisfaction with time to comprehension</td><td align="left" valign="top">&#x003C;.001</td></tr><tr><td align="left" valign="top">&#x2003;Very satisfied/somewhat satisfied</td><td align="left" valign="top">27 (23)</td><td align="left" valign="top">115 (97)</td><td align="left" valign="top"/></tr><tr><td align="left" valign="top">&#x2003;Neutral</td><td align="left" valign="top">31 (26)</td><td align="left" valign="top">2 (2)</td><td align="left" valign="top"/></tr><tr><td align="left" valign="top">&#x2003;Very dissatisfied/somewhat dissatisfied</td><td align="left" valign="top">60 (51)</td><td align="left" valign="top">1 (1)</td><td align="left" valign="top"/></tr></tbody></table></table-wrap><p>The summaries demonstrated high utility, with 92% (108/118) of responses indicating they clarified questions about the original report and improved patients&#x2019; understanding of clinical implications. When asked about format preferences, 67% (79/118) of responses indicated a preference for having access to both the summary and original report, while 26% (31/118) of responses preferred to have the summary alone. Only 1% (1/118) preferred the original report exclusively, with the remaining 6% (7/120) indicating a preference for alternative formats.</p><p>Qualitative analysis of patient feedback revealed several themes regarding desired improvements to the summaries. Of the free-text responses, 50% (53/107) indicated no improvements were needed. The remaining responses suggested enhancing the professional tone (13%, 14/107), providing more specific follow-up recommendations (14%, 15/107), and including additional detail about medical terminology and physiological processes relevant to the findings (23%, 25/107).</p></sec><sec id="s3-4"><title>Effect of the Summary on Patient Understanding</title><p>Patient understanding was assessed by comparing each patient&#x2019;s identification of the main point of the report/summary against the radiologist&#x2019;s response regarding the main point of the report. When reading the original radiologist report, 60% (71/119) of patient responses aligned with radiologist interpretations, while 40% (48/119) were incongruent. Patient understanding improved with the summary, with 74% (88/119) of responses aligning with radiologist interpretations, while 27% (32/119) remained incongruent.</p><p>When comparing patient responses regarding the main point the radiologist was trying to convey after reading only the report with their response after reading the summary, answers did not change in 52% (62/119) (ie, responses were congruent). In 29% (34/119), the responses after reading the summary were different (ie, the summary changed a patient&#x2019;s understanding). Specifically, in 23% (27/119), responses after reading the summary contained additional relevant details or demonstrated closer alignment with radiologists&#x2019; responses, suggesting an increase in understanding as a result of the summary. In 6% (7/119), responses after reading the summary contained fewer details or diverged further from radiologists&#x2019; responses, suggesting decreased understanding as a result of the summary.</p></sec></sec><sec id="s4" sec-type="discussion"><title>Discussion</title><sec id="s4-1"><title>Principal Findings</title><p>This study demonstrates that LLMs can successfully generate neuroradiology reports that may be more understandable to patients while maintaining clinical accuracy. The AI-generated summaries significantly improved readability metrics across all measured scales, reducing the required reading level from college (grade 12.5) to middle school (grade 7.2), while substantially increasing reading ease scores [<xref ref-type="bibr" rid="ref25">25</xref>,<xref ref-type="bibr" rid="ref26">26</xref>]. This improvement in accessibility was achieved while maintaining high accuracy, with radiologists confirming that 83% (100/120) of summaries represented the original findings &#x201C;extremely&#x201D; or &#x201C;very well.&#x201D; Similar studies have shown that such summaries maintain high accuracy [<xref ref-type="bibr" rid="ref27">27</xref>,<xref ref-type="bibr" rid="ref28">28</xref>].</p><p>A unique strength of this study is the inclusion of patients in the evaluation of reports and AI-generated summaries, as the vast majority of research focuses on evaluation by radiologists and nonradiologist physicians [<xref ref-type="bibr" rid="ref2">2</xref>,<xref ref-type="bibr" rid="ref29">29</xref>-<xref ref-type="bibr" rid="ref31">31</xref>]. Patient response data revealed a striking contrast between comprehension of traditional reports and AI-generated summaries, with only 26% (31/118) of patients reporting confidence in understanding the original reports and 98% (116/118) reporting confidence in understanding the summaries. In addition, the high satisfaction rates with the summaries&#x2019; level of detail (98%, 116/118) and medical terminology (91%, 107/118) indicate that the AI successfully struck a balance between accessibility and informativeness.</p><p>The dramatic improvement in patient confidence was accompanied by objective improvements in understanding, with congruence between patient and radiologist interpretations increasing from 60% (71/119) on the original reports to 74% (88/119) after reading the summaries. In addition, 23% (27/119) of responses after reading the summary contained additional relevant details or demonstrated closer alignment with radiologist responses, consistent with an increased understanding as a result of the summary. This suggests that AI-generated summaries not only increase patient confidence but also lead to a more accurate understanding of radiological findings. This aligns with prior evidence that patient-friendly summaries can enhance patient understanding; for example, nonphysician readers in one study had comprehension scores of 4.69/6 for AI-generated summaries versus 2.71/5 for original reports [<xref ref-type="bibr" rid="ref27">27</xref>].</p><p>Particularly noteworthy was the strong patient preference for having access to both the original report and the summary (67%, 79/118 of responses), with 98% (116/118) of responses indicating they would use the summary if available. This suggests that patients value the combination of detailed technical information with accessible explanations, rather than seeing the summary as a replacement for the full report. These findings align with the shifting preference among the general public toward greater transparency in health care [<xref ref-type="bibr" rid="ref3">3</xref>,<xref ref-type="bibr" rid="ref4">4</xref>,<xref ref-type="bibr" rid="ref7">7</xref>,<xref ref-type="bibr" rid="ref8">8</xref>].</p><p>Importantly, no frank hallucinations were observed in the summaries, which is crucial for clinical implementation. However, a small percentage (10%, 12/120) of radiologist responses indicated that summaries may be misleading. Radiologists noted instances of omission and both over- and underemphasis of findings (12%, 14/120 and 18%, 22/120 of responses, respectively), particularly regarding incidental findings and critical abnormalities. The concern about misleading summaries was corroborated by the 6% (7/119) of patient responses that contained fewer details or diverged further from radiologist responses after reading the summary. Misleading information has the potential to lead to a variety of negative consequences for the patient, further emphasizing the need for comprehensive standards surrounding LLM-generated resources [<xref ref-type="bibr" rid="ref20">20</xref>].</p><p>In addition, in some instances, language was deemed too casual by radiologists (eg, &#x201C;squished&#x201D; instead of &#x201C;compressed&#x201D;), highlighting the need for more sophisticated prompt engineering to maintain professional medical communication standards while improving accessibility. An area of improvement identified by the patient reviewers was highlighted by the 14% (15/107) of patient responses indicating a desire for more explicit guidance on next steps and implications of findings.</p></sec><sec id="s4-2"><title>Limitations</title><p>This study has several limitations. Principally, its focus on neuroradiology reports from a single institution potentially limits the generalizability of our findings to other radiologic subspecialties and other institutional reporting patterns. In addition, the study included a relatively small number of patient and radiologist reviewers. We hope that this work encourages other investigators to replicate this study with larger cohorts of reviewers to assess its applicability in subspecialties outside of neuroradiology and in larger and more diverse patient populations. Patient volunteers were relatively older and had a relatively higher average education level than the general population and were members of our health institution&#x2019;s Patient Family Advisory Council. This likely indicates a higher level of health literacy and reading ability compared to the average patient. Overall, we anticipate that if anything, this biased the results toward greater understanding of the original radiology report, rather than less. Finally, volunteers were not reading their own reports, which removes the emotional salience and may alter the experience.</p></sec><sec id="s4-3"><title>Considerations for Clinical Deployment</title><p>Clinical implementation of AI-generated summaries of radiology reports raises a number of ethical and practical considerations that must be addressed. In reference to ChatGPT, OpenAI policy explicitly requires that AI-generated medical content undergo review by qualified professionals due to the risk of AI errors, which would necessitate radiologist or ordering provider oversight of all summaries before patients can access them [<xref ref-type="bibr" rid="ref32">32</xref>]. The choice of model is incredibly important, as different models perform differently on such summarization tasks [<xref ref-type="bibr" rid="ref28">28</xref>]. Clinical deployment requires HIPAA (Health Insurance Portability and Accountability Act)-compliant platforms (many are already available) and integration into existing electronic health record workflows without excessive burden on providers. Given the extraordinary burden already placed on radiologists to complete standard dictations amid increasing imaging volumes, adding review of an AI-generated summary to the workflow after final signature may not be feasible. It is important to note, however, that patients are already using consumer AI tools/chatbots to interpret their medical records without professional oversight, potentially receiving inaccurate or anxiety-provoking information [<xref ref-type="bibr" rid="ref33">33</xref>]. Institution-supervised AI summaries could provide a safer alternative to this unregulated landscape while ensuring some degree of professional oversight and quality assurance. Future implementations must balance the demonstrated benefits of improved patient comprehension with robust safeguards.</p></sec><sec id="s4-4"><title>Future Research</title><p>AI-generated, patient-friendly summaries appear to represent a promising tool for improving patient understanding of radiology reports. The summaries had high accuracy rates and dramatically improved readability and patient comprehension while maintaining clinical precision. The perceived benefit to patients in conjunction with the objective increase in comprehension suggests potential benefit in terms of patient experience and outcomes, though further investigation is needed using patients&#x2019; own scans and reports (which we are pursuing) [<xref ref-type="bibr" rid="ref1">1</xref>,<xref ref-type="bibr" rid="ref3">3</xref>,<xref ref-type="bibr" rid="ref8">8</xref>,<xref ref-type="bibr" rid="ref10">10</xref>-<xref ref-type="bibr" rid="ref12">12</xref>]. Future work should also investigate how prompt optimization and custom model development may improve the quality of outputs. Integration of these summaries into existing radiology workflows and patient portals should be studied, with particular attention to the ultimate impact on radiologist workload, patient experience, and outcomes. Conscientious integration of this technology is required, given the extraction of sensitive patient data and associated concerns [<xref ref-type="bibr" rid="ref34">34</xref>].</p><p>Future research should explore options to continue refining the clinical accuracy of these reports. Though LLMs lack the ability to independently reason, there are several ways to optimize their output, including optimization of prompt engineering, incorporating structured templates for critical findings, and using standardized language for common pathologies [<xref ref-type="bibr" rid="ref2">2</xref>,<xref ref-type="bibr" rid="ref30">30</xref>,<xref ref-type="bibr" rid="ref31">31</xref>]. In addition, providing the LLM with more clinical context (eg, recent note from an ordering provider, problem list, patient class [IP/OP/ED]) may improve the quality of the summary and recommendations provided [<xref ref-type="bibr" rid="ref31">31</xref>,<xref ref-type="bibr" rid="ref35">35</xref>].</p></sec><sec id="s4-5"><title>Conclusions</title><p>This study demonstrates that ChatGPT-generated summaries can help bridge the communication gap between complex radiology reports and patient understanding while maintaining accuracy. While promising, our findings highlight the need for continued refinement to address instances of overemphasis or underemphasis, which can mislead patients in certain circumstances. As such, it is critically important to establish robust quality assurance frameworks before clinical implementation.</p></sec></sec></body><back><notes><sec><title>Data Availability</title><p>The datasets generalized and analyzed during this study are available from the corresponding author on reasonable request.</p></sec></notes><fn-group><fn fn-type="conflict"><p>None declared.</p></fn></fn-group><glossary><title>Abbreviations</title><def-list><def-item><term id="abb1">AI</term><def><p>artificial intelligence</p></def></def-item><def-item><term id="abb2">HIPAA</term><def><p>Health Insurance Portability and Accountability Act</p></def></def-item><def-item><term id="abb3">LLM</term><def><p>large language model</p></def></def-item><def-item><term id="abb4">SMOG</term><def><p>Simple Measure of Gobbledygook</p></def></def-item></def-list></glossary><ref-list><title>References</title><ref id="ref1"><label>1</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Amin</surname><given-names>K</given-names> </name><name name-style="western"><surname>Khosla</surname><given-names>P</given-names> </name><name name-style="western"><surname>Doshi</surname><given-names>R</given-names> </name><name name-style="western"><surname>Chheang</surname><given-names>S</given-names> </name><name name-style="western"><surname>Forman</surname><given-names>HP</given-names> </name></person-group><article-title>Artificial intelligence to improve patient understanding of radiology reports</article-title><source>Yale J Biol Med</source><year>2023</year><month>09</month><volume>96</volume><issue>3</issue><fpage>407</fpage><lpage>417</lpage><pub-id pub-id-type="doi">10.59249/NKOY5498</pub-id><pub-id pub-id-type="medline">37780992</pub-id></nlm-citation></ref><ref id="ref2"><label>2</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Doshi</surname><given-names>R</given-names> </name><name name-style="western"><surname>Amin</surname><given-names>KS</given-names> </name><name name-style="western"><surname>Khosla</surname><given-names>P</given-names> </name><name name-style="western"><surname>Bajaj</surname><given-names>SS</given-names> </name><name name-style="western"><surname>Chheang</surname><given-names>S</given-names> </name><name name-style="western"><surname>Forman</surname><given-names>HP</given-names> </name></person-group><article-title>Quantitative evaluation of large language models to streamline radiology report impressions: a multimodal retrospective analysis</article-title><source>Radiology</source><year>2024</year><month>03</month><volume>310</volume><issue>3</issue><fpage>e231593</fpage><pub-id pub-id-type="doi">10.1148/radiol.231593</pub-id><pub-id pub-id-type="medline">38530171</pub-id></nlm-citation></ref><ref id="ref3"><label>3</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Alarifi</surname><given-names>M</given-names> </name><name name-style="western"><surname>Patrick</surname><given-names>T</given-names> </name><name name-style="western"><surname>Jabour</surname><given-names>A</given-names> </name><name name-style="western"><surname>Wu</surname><given-names>M</given-names> </name><name name-style="western"><surname>Luo</surname><given-names>J</given-names> </name></person-group><article-title>Full radiology report through patient web portal: a literature review</article-title><source>Int J Environ Res Public Health</source><year>2020</year><month>05</month><day>22</day><volume>17</volume><issue>10</issue><fpage>3673</fpage><pub-id pub-id-type="doi">10.3390/ijerph17103673</pub-id><pub-id pub-id-type="medline">32456099</pub-id></nlm-citation></ref><ref id="ref4"><label>4</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Lee</surname><given-names>CI</given-names> </name><name name-style="western"><surname>Langlotz</surname><given-names>CP</given-names> </name><name name-style="western"><surname>Elmore</surname><given-names>JG</given-names> </name></person-group><article-title>Implications of direct patient online access to radiology reports through patient web portals</article-title><source>J Am Coll Radiol</source><year>2016</year><month>12</month><volume>13</volume><issue>12 Pt B</issue><fpage>1608</fpage><lpage>1614</lpage><pub-id pub-id-type="doi">10.1016/j.jacr.2016.09.007</pub-id><pub-id pub-id-type="medline">27888949</pub-id></nlm-citation></ref><ref id="ref5"><label>5</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Halaska</surname><given-names>C</given-names> </name><name name-style="western"><surname>Sachs</surname><given-names>P</given-names> </name><name name-style="western"><surname>Sanfilippo</surname><given-names>K</given-names> </name><name name-style="western"><surname>Lin</surname><given-names>CT</given-names> </name></person-group><article-title>Patient attitudes about viewing their radiology images online: preintervention survey</article-title><source>J Med Internet Res</source><year>2019</year><month>07</month><day>18</day><volume>21</volume><issue>7</issue><fpage>e12595</fpage><pub-id pub-id-type="doi">10.2196/12595</pub-id><pub-id pub-id-type="medline">31322124</pub-id></nlm-citation></ref><ref id="ref6"><label>6</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Kemp</surname><given-names>J</given-names> </name><name name-style="western"><surname>Short</surname><given-names>R</given-names> </name><name name-style="western"><surname>Bryant</surname><given-names>S</given-names> </name><name name-style="western"><surname>Sample</surname><given-names>L</given-names> </name><name name-style="western"><surname>Befera</surname><given-names>N</given-names> </name></person-group><article-title>Patient-friendly radiology reporting-implementation and outcomes</article-title><source>J Am Coll Radiol</source><year>2022</year><month>02</month><volume>19</volume><issue>2 Pt B</issue><fpage>377</fpage><lpage>383</lpage><pub-id pub-id-type="doi">10.1016/j.jacr.2021.10.008</pub-id><pub-id pub-id-type="medline">35152963</pub-id></nlm-citation></ref><ref id="ref7"><label>7</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Farmer</surname><given-names>CI</given-names> </name><name name-style="western"><surname>Bourne</surname><given-names>AM</given-names> </name><name name-style="western"><surname>O&#x2019;Connor</surname><given-names>D</given-names> </name><name name-style="western"><surname>Jarvik</surname><given-names>JG</given-names> </name><name name-style="western"><surname>Buchbinder</surname><given-names>R</given-names> </name></person-group><article-title>Enhancing clinician and patient understanding of radiology reports: a scoping review of international guidelines</article-title><source>Insights Imaging</source><year>2020</year><month>05</month><day>5</day><volume>11</volume><issue>1</issue><fpage>62</fpage><pub-id pub-id-type="doi">10.1186/s13244-020-00864-9</pub-id><pub-id pub-id-type="medline">32372369</pub-id></nlm-citation></ref><ref id="ref8"><label>8</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Hansberry</surname><given-names>DR</given-names> </name><name name-style="western"><surname>Donovan</surname><given-names>AL</given-names> </name><name name-style="western"><surname>Prabhu</surname><given-names>AV</given-names> </name><name name-style="western"><surname>Agarwal</surname><given-names>N</given-names> </name><name name-style="western"><surname>Cox</surname><given-names>M</given-names> </name><name name-style="western"><surname>Flanders</surname><given-names>AE</given-names> </name></person-group><article-title>Enhancing the radiologist-patient relationship through improved communication: a quantitative readability analysis in spine radiology</article-title><source>AJNR Am J Neuroradiol</source><year>2017</year><month>06</month><volume>38</volume><issue>6</issue><fpage>1252</fpage><lpage>1256</lpage><pub-id pub-id-type="doi">10.3174/ajnr.A5151</pub-id><pub-id pub-id-type="medline">28385883</pub-id></nlm-citation></ref><ref id="ref9"><label>9</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Yi</surname><given-names>PH</given-names> </name><name name-style="western"><surname>Golden</surname><given-names>SK</given-names> </name><name name-style="western"><surname>Harringa</surname><given-names>JB</given-names> </name><name name-style="western"><surname>Kliewer</surname><given-names>MA</given-names> </name></person-group><article-title>Readability of lumbar spine MRI reports: will patients understand?</article-title><source>AJR Am J Roentgenol</source><year>2019</year><month>03</month><volume>212</volume><issue>3</issue><fpage>602</fpage><lpage>606</lpage><pub-id pub-id-type="doi">10.2214/AJR.18.20197</pub-id><pub-id pub-id-type="medline">30620671</pub-id></nlm-citation></ref><ref id="ref10"><label>10</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Johnson</surname><given-names>AM</given-names> </name><name name-style="western"><surname>Brimhall</surname><given-names>AS</given-names> </name><name name-style="western"><surname>Johnson</surname><given-names>ET</given-names> </name><etal/></person-group><article-title>A systematic review of the effectiveness of patient education through patient portals</article-title><source>JAMIA Open</source><year>2023</year><month>04</month><volume>6</volume><issue>1</issue><fpage>ooac085</fpage><pub-id pub-id-type="doi">10.1093/jamiaopen/ooac085</pub-id><pub-id pub-id-type="medline">36686972</pub-id></nlm-citation></ref><ref id="ref11"><label>11</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Alturkistani</surname><given-names>A</given-names> </name><name name-style="western"><surname>Qavi</surname><given-names>A</given-names> </name><name name-style="western"><surname>Anyanwu</surname><given-names>PE</given-names> </name><name name-style="western"><surname>Greenfield</surname><given-names>G</given-names> </name><name name-style="western"><surname>Greaves</surname><given-names>F</given-names> </name><name name-style="western"><surname>Costelloe</surname><given-names>C</given-names> </name></person-group><article-title>Patient portal functionalities and patient outcomes among patients with diabetes: systematic review</article-title><source>J Med Internet Res</source><year>2020</year><month>09</month><day>22</day><volume>22</volume><issue>9</issue><fpage>e18976</fpage><pub-id pub-id-type="doi">10.2196/18976</pub-id><pub-id pub-id-type="medline">32960182</pub-id></nlm-citation></ref><ref id="ref12"><label>12</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Bange</surname><given-names>M</given-names> </name><name name-style="western"><surname>Huh</surname><given-names>E</given-names> </name><name name-style="western"><surname>Novin</surname><given-names>SA</given-names> </name><name name-style="western"><surname>Hui</surname><given-names>FK</given-names> </name><name name-style="western"><surname>Yi</surname><given-names>PH</given-names> </name></person-group><article-title>Readability of patient education materials from RadiologyInfo.org: has there been progress over the past 5 years?</article-title><source>AJR Am J Roentgenol</source><year>2019</year><month>10</month><volume>213</volume><issue>4</issue><fpage>875</fpage><lpage>879</lpage><pub-id pub-id-type="doi">10.2214/AJR.18.21047</pub-id><pub-id pub-id-type="medline">31386570</pub-id></nlm-citation></ref><ref id="ref13"><label>13</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Qenam</surname><given-names>B</given-names> </name><name name-style="western"><surname>Kim</surname><given-names>TY</given-names> </name><name name-style="western"><surname>Carroll</surname><given-names>MJ</given-names> </name><name name-style="western"><surname>Hogarth</surname><given-names>M</given-names> </name></person-group><article-title>Text simplification using consumer health vocabulary to generate patient-centered radiology reporting: translation and evaluation</article-title><source>J Med Internet Res</source><year>2017</year><month>12</month><day>18</day><volume>19</volume><issue>12</issue><fpage>e417</fpage><pub-id pub-id-type="doi">10.2196/jmir.8536</pub-id><pub-id pub-id-type="medline">29254915</pub-id></nlm-citation></ref><ref id="ref14"><label>14</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Schwartz</surname><given-names>LH</given-names> </name><name name-style="western"><surname>Panicek</surname><given-names>DM</given-names> </name><name name-style="western"><surname>Berk</surname><given-names>AR</given-names> </name><name name-style="western"><surname>Li</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Hricak</surname><given-names>H</given-names> </name></person-group><article-title>Improving communication of diagnostic radiology findings through structured reporting</article-title><source>Radiology</source><year>2011</year><month>07</month><volume>260</volume><issue>1</issue><fpage>174</fpage><lpage>181</lpage><pub-id pub-id-type="doi">10.1148/radiol.11101913</pub-id><pub-id pub-id-type="medline">21518775</pub-id></nlm-citation></ref><ref id="ref15"><label>15</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Norris</surname><given-names>EC</given-names> </name><name name-style="western"><surname>Halaska</surname><given-names>C</given-names> </name><name name-style="western"><surname>Sachs</surname><given-names>PB</given-names> </name><name name-style="western"><surname>Lin</surname><given-names>CT</given-names> </name><name name-style="western"><surname>Sanfilippo</surname><given-names>K</given-names> </name><name name-style="western"><surname>Honce</surname><given-names>JM</given-names> </name></person-group><article-title>Understanding patient experiences, opinions, and actions taken after viewing their own radiology images online: web-based survey</article-title><source>JMIR Form Res</source><year>2022</year><month>04</month><day>25</day><volume>6</volume><issue>4</issue><fpage>e29496</fpage><pub-id pub-id-type="doi">10.2196/29496</pub-id><pub-id pub-id-type="medline">35468086</pub-id></nlm-citation></ref><ref id="ref16"><label>16</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>McEnteggart</surname><given-names>GE</given-names> </name><name name-style="western"><surname>Naeem</surname><given-names>M</given-names> </name><name name-style="western"><surname>Skierkowski</surname><given-names>D</given-names> </name><name name-style="western"><surname>Baird</surname><given-names>GL</given-names> </name><name name-style="western"><surname>Ahn</surname><given-names>SH</given-names> </name><name name-style="western"><surname>Soares</surname><given-names>G</given-names> </name></person-group><article-title>Readability of online patient education materials related to IR</article-title><source>J Vasc Interv Radiol</source><year>2015</year><month>08</month><volume>26</volume><issue>8</issue><fpage>1164</fpage><lpage>1168</lpage><pub-id pub-id-type="doi">10.1016/j.jvir.2015.03.019</pub-id><pub-id pub-id-type="medline">25935147</pub-id></nlm-citation></ref><ref id="ref17"><label>17</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Recht</surname><given-names>MP</given-names> </name><name name-style="western"><surname>Donoso-Bach</surname><given-names>L</given-names> </name><name name-style="western"><surname>Brklja&#x010D;i&#x0107;</surname><given-names>B</given-names> </name><name name-style="western"><surname>Chandarana</surname><given-names>H</given-names> </name><name name-style="western"><surname>Jankharia</surname><given-names>B</given-names> </name><name name-style="western"><surname>Mahoney</surname><given-names>MC</given-names> </name></person-group><article-title>Patient-centered radiology: a roadmap for outpatient imaging</article-title><source>Eur Radiol</source><year>2024</year><month>07</month><volume>34</volume><issue>7</issue><fpage>4331</fpage><lpage>4340</lpage><pub-id pub-id-type="doi">10.1007/s00330-023-10370-3</pub-id><pub-id pub-id-type="medline">38047974</pub-id></nlm-citation></ref><ref id="ref18"><label>18</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Liebrenz</surname><given-names>M</given-names> </name><name name-style="western"><surname>Schleifer</surname><given-names>R</given-names> </name><name name-style="western"><surname>Buadze</surname><given-names>A</given-names> </name><name name-style="western"><surname>Bhugra</surname><given-names>D</given-names> </name><name name-style="western"><surname>Smith</surname><given-names>A</given-names> </name></person-group><article-title>Generating scholarly content with ChatGPT: ethical challenges for medical publishing</article-title><source>Lancet Digit Health</source><year>2023</year><month>03</month><volume>5</volume><issue>3</issue><fpage>e105</fpage><lpage>e106</lpage><pub-id pub-id-type="doi">10.1016/S2589-7500(23)00019-5</pub-id><pub-id pub-id-type="medline">36754725</pub-id></nlm-citation></ref><ref id="ref19"><label>19</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Yeo</surname><given-names>YH</given-names> </name><name name-style="western"><surname>Samaan</surname><given-names>JS</given-names> </name><name name-style="western"><surname>Ng</surname><given-names>WH</given-names> </name><etal/></person-group><article-title>Assessing the performance of ChatGPT in answering questions regarding cirrhosis and hepatocellular carcinoma</article-title><source>Clin Mol Hepatol</source><year>2023</year><month>07</month><volume>29</volume><issue>3</issue><fpage>721</fpage><lpage>732</lpage><pub-id pub-id-type="doi">10.3350/cmh.2023.0089</pub-id><pub-id pub-id-type="medline">35964695</pub-id></nlm-citation></ref><ref id="ref20"><label>20</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Goodman</surname><given-names>KE</given-names> </name><name name-style="western"><surname>Yi</surname><given-names>PH</given-names> </name><name name-style="western"><surname>Morgan</surname><given-names>DJ</given-names> </name></person-group><article-title>AI-generated clinical summaries require more than accuracy</article-title><source>JAMA</source><year>2024</year><month>02</month><day>27</day><volume>331</volume><issue>8</issue><fpage>637</fpage><lpage>638</lpage><pub-id pub-id-type="doi">10.1001/jama.2024.0555</pub-id><pub-id pub-id-type="medline">38285439</pub-id></nlm-citation></ref><ref id="ref21"><label>21</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Parillo</surname><given-names>M</given-names> </name><name name-style="western"><surname>Vaccarino</surname><given-names>F</given-names> </name><name name-style="western"><surname>Beomonte Zobel</surname><given-names>B</given-names> </name><name name-style="western"><surname>Mallio</surname><given-names>CA</given-names> </name></person-group><article-title>ChatGPT and radiology report: potential applications and limitations</article-title><source>Radiol med</source><volume>129</volume><issue>12</issue><fpage>1849</fpage><lpage>1863</lpage><pub-id pub-id-type="doi">10.1007/s11547-024-01915-7</pub-id></nlm-citation></ref><ref id="ref22"><label>22</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Wang</surname><given-names>LW</given-names> </name><name name-style="western"><surname>Miller</surname><given-names>MJ</given-names> </name><name name-style="western"><surname>Schmitt</surname><given-names>MR</given-names> </name><name name-style="western"><surname>Wen</surname><given-names>FK</given-names> </name></person-group><article-title>Assessing readability formula differences with written health information materials: application, results, and recommendations</article-title><source>Res Social Adm Pharm</source><year>2013</year><volume>9</volume><issue>5</issue><fpage>503</fpage><lpage>516</lpage><pub-id pub-id-type="doi">10.1016/j.sapharm.2012.05.009</pub-id><pub-id pub-id-type="medline">22835706</pub-id></nlm-citation></ref><ref id="ref23"><label>23</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Flesch</surname><given-names>R</given-names> </name></person-group><article-title>A new readability yardstick</article-title><source>Journal of Applied Psychology</source><year>1948</year><month>06</month><volume>32</volume><issue>3</issue><fpage>221</fpage><lpage>233</lpage><pub-id pub-id-type="doi">10.1037/h0057532</pub-id><pub-id pub-id-type="medline">18867058</pub-id></nlm-citation></ref><ref id="ref24"><label>24</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Sabharwal</surname><given-names>S</given-names> </name><name name-style="western"><surname>Badarudeen</surname><given-names>S</given-names> </name><name name-style="western"><surname>Unes Kunju</surname><given-names>S</given-names> </name></person-group><article-title>Readability of online patient education materials from the AAOS web site</article-title><source>Clin Orthop Relat Res</source><year>2008</year><month>05</month><volume>466</volume><issue>5</issue><fpage>1245</fpage><lpage>1250</lpage><pub-id pub-id-type="doi">10.1007/s11999-008-0193-8</pub-id><pub-id pub-id-type="medline">18324452</pub-id></nlm-citation></ref><ref id="ref25"><label>25</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Haver</surname><given-names>HL</given-names> </name><name name-style="western"><surname>Gupta</surname><given-names>AK</given-names> </name><name name-style="western"><surname>Ambinder</surname><given-names>EB</given-names> </name><etal/></person-group><article-title>Evaluating the use of ChatGPT to accurately simplify patient-centered information about breast cancer prevention and screening</article-title><source>Radiol Imaging Cancer</source><year>2024</year><month>03</month><volume>6</volume><issue>2</issue><fpage>e230086</fpage><pub-id pub-id-type="doi">10.1148/rycan.230086</pub-id><pub-id pub-id-type="medline">38305716</pub-id></nlm-citation></ref><ref id="ref26"><label>26</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Chung</surname><given-names>EM</given-names> </name><name name-style="western"><surname>Zhang</surname><given-names>SC</given-names> </name><name name-style="western"><surname>Nguyen</surname><given-names>AT</given-names> </name><name name-style="western"><surname>Atkins</surname><given-names>KM</given-names> </name><name name-style="western"><surname>Sandler</surname><given-names>HM</given-names> </name><name name-style="western"><surname>Kamrava</surname><given-names>M</given-names> </name></person-group><article-title>Feasibility and acceptability of ChatGPT generated radiology report summaries for cancer patients</article-title><source>Digit Health</source><year>2023</year><volume>9</volume><fpage>20552076231221620</fpage><pub-id pub-id-type="doi">10.1177/20552076231221620</pub-id><pub-id pub-id-type="medline">38130802</pub-id></nlm-citation></ref><ref id="ref27"><label>27</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Park</surname><given-names>J</given-names> </name><name name-style="western"><surname>Oh</surname><given-names>K</given-names> </name><name name-style="western"><surname>Han</surname><given-names>K</given-names> </name><name name-style="western"><surname>Lee</surname><given-names>YH</given-names> </name></person-group><article-title>Patient-centered radiology reports with generative artificial intelligence: adding value to radiology reporting</article-title><source>Sci Rep</source><year>2024</year><volume>14</volume><issue>1</issue><fpage>13218</fpage><pub-id pub-id-type="doi">10.1038/s41598-024-63824-z</pub-id></nlm-citation></ref><ref id="ref28"><label>28</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Tepe</surname><given-names>M</given-names> </name><name name-style="western"><surname>Emekli</surname><given-names>E</given-names> </name></person-group><article-title>Decoding medical jargon: the use of AI language models (ChatGPT-4, BARD, microsoft copilot) in radiology reports</article-title><source>Patient Educ Couns</source><year>2024</year><month>09</month><volume>126</volume><fpage>108307</fpage><pub-id pub-id-type="doi">10.1016/j.pec.2024.108307</pub-id><pub-id pub-id-type="medline">38743965</pub-id></nlm-citation></ref><ref id="ref29"><label>29</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Kuckelman</surname><given-names>IJ</given-names> </name><name name-style="western"><surname>Wetley</surname><given-names>K</given-names> </name><name name-style="western"><surname>Yi</surname><given-names>PH</given-names> </name><name name-style="western"><surname>Ross</surname><given-names>AB</given-names> </name></person-group><article-title>Translating musculoskeletal radiology reports into patient-friendly summaries using ChatGPT-4</article-title><source>Skeletal Radiol</source><year>2024</year><month>08</month><volume>53</volume><issue>8</issue><fpage>1621</fpage><lpage>1624</lpage><pub-id pub-id-type="doi">10.1007/s00256-024-04599-2</pub-id><pub-id pub-id-type="medline">38270616</pub-id></nlm-citation></ref><ref id="ref30"><label>30</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Jeblick</surname><given-names>K</given-names> </name><name name-style="western"><surname>Schachtner</surname><given-names>B</given-names> </name><name name-style="western"><surname>Dexl</surname><given-names>J</given-names> </name><etal/></person-group><article-title>ChatGPT makes medicine easy to swallow: an exploratory case study on simplified radiology reports</article-title><source>Eur Radiol</source><year>2024</year><month>05</month><volume>34</volume><issue>5</issue><fpage>2817</fpage><lpage>2825</lpage><pub-id pub-id-type="doi">10.1007/s00330-023-10213-1</pub-id><pub-id pub-id-type="medline">37794249</pub-id></nlm-citation></ref><ref id="ref31"><label>31</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Lyu</surname><given-names>Q</given-names> </name><name name-style="western"><surname>Tan</surname><given-names>J</given-names> </name><name name-style="western"><surname>Zapadka</surname><given-names>ME</given-names> </name><etal/></person-group><article-title>Translating radiology reports into plain language using ChatGPT and GPT-4 with prompt learning: results, limitations, and potential</article-title><source>Vis Comput Ind Biomed Art</source><year>2023</year><month>05</month><day>18</day><volume>6</volume><issue>1</issue><fpage>9</fpage><pub-id pub-id-type="doi">10.1186/s42492-023-00136-5</pub-id><pub-id pub-id-type="medline">37198498</pub-id></nlm-citation></ref><ref id="ref32"><label>32</label><nlm-citation citation-type="web"><article-title>Usage policies</article-title><source>OpenAI</source><access-date>2025-06-20</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://openai.com/policies/usage-policies/">https://openai.com/policies/usage-policies/</ext-link></comment></nlm-citation></ref><ref id="ref33"><label>33</label><nlm-citation citation-type="web"><article-title>Consumer perceptions of AI in healthcare</article-title><source>UserTesting</source><access-date>2025-06-20</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://www.usertesting.com/resources/reports/consumer-perceptions-ai-healthcare">https://www.usertesting.com/resources/reports/consumer-perceptions-ai-healthcare</ext-link></comment></nlm-citation></ref><ref id="ref34"><label>34</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Patel</surname><given-names>SB</given-names> </name><name name-style="western"><surname>Lam</surname><given-names>K</given-names> </name></person-group><article-title>ChatGPT: the future of discharge summaries?</article-title><source>Lancet Digit Health</source><year>2023</year><month>03</month><volume>5</volume><issue>3</issue><fpage>e107</fpage><lpage>e108</lpage><pub-id pub-id-type="doi">10.1016/S2589-7500(23)00021-3</pub-id><pub-id pub-id-type="medline">36754724</pub-id></nlm-citation></ref><ref id="ref35"><label>35</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Campbell</surname><given-names>DJ</given-names> </name><name name-style="western"><surname>Estephan</surname><given-names>LE</given-names> </name></person-group><article-title>ChatGPT, obstructive sleep apnea, and patient education</article-title><source>J Clin Sleep Med</source><year>2023</year><access-date>2025-06-20</access-date><volume>19</volume><issue>12</issue><fpage>2133</fpage><comment><ext-link ext-link-type="uri" xlink:href="https://jcsm.aasm.org/doi/abs/10.5664/jcsm.10768">https://jcsm.aasm.org/doi/abs/10.5664/jcsm.10768</ext-link></comment><pub-id pub-id-type="doi">10.5664/jcsm.10808</pub-id></nlm-citation></ref></ref-list><app-group><supplementary-material id="app1"><label>Multimedia Appendix 1</label><p>Complete patient and radiology survey instruments.</p><media xlink:href="formative_v9i1e76097_app1.docx" xlink:title="DOCX File, 22 KB"/></supplementary-material></app-group></back></article>