<?xml version="1.0" encoding="UTF-8"?><!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.0 20040830//EN" "journalpublishing.dtd"><article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" dtd-version="2.0" xml:lang="en" article-type="research-article"><front><journal-meta><journal-id journal-id-type="nlm-ta">JMIR Form Res</journal-id><journal-id journal-id-type="publisher-id">formative</journal-id><journal-id journal-id-type="index">27</journal-id><journal-title>JMIR Formative Research</journal-title><abbrev-journal-title>JMIR Form Res</abbrev-journal-title><issn pub-type="epub">2561-326X</issn><publisher><publisher-name>JMIR Publications</publisher-name><publisher-loc>Toronto, Canada</publisher-loc></publisher></journal-meta><article-meta><article-id pub-id-type="publisher-id">v10i1e90110</article-id><article-id pub-id-type="doi">10.2196/90110</article-id><article-categories><subj-group subj-group-type="heading"><subject>Original Paper</subject></subj-group></article-categories><title-group><article-title>Development of a Child Articulation Screening Test Within Digital Therapeutics: Delphi Study</article-title></title-group><contrib-group><contrib contrib-type="author"><name name-style="western"><surname>Lee</surname><given-names>Jaewon</given-names></name><degrees>MSc, PT</degrees><xref ref-type="aff" rid="aff1">1</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Kwon</surname><given-names>Sejin</given-names></name><degrees>BHS, PT</degrees><xref ref-type="aff" rid="aff1">1</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Ko</surname><given-names>Jin Young</given-names></name><degrees>MD</degrees><xref ref-type="aff" rid="aff2">2</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Park</surname><given-names>Yulhyun</given-names></name><degrees>MD</degrees><xref ref-type="aff" rid="aff3">3</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Lee</surname><given-names>Jinju</given-names></name><degrees>MSc, PT</degrees><xref ref-type="aff" rid="aff1">1</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Ryu</surname><given-names>Ju Seok</given-names></name><degrees>MD, PhD</degrees><xref ref-type="aff" rid="aff1">1</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Yoon</surname><given-names>Seo Yeon</given-names></name><degrees>MD, PhD</degrees><xref ref-type="aff" rid="aff4">4</xref></contrib><contrib contrib-type="author" corresp="yes"><name name-style="western"><surname>Suh</surname><given-names>Jee Hyun</given-names></name><degrees>MD, PhD</degrees><xref ref-type="aff" rid="aff1">1</xref></contrib></contrib-group><aff id="aff1"><institution>Department of Rehabilitation Medicine, Seoul National University Bundang Hospital</institution><addr-line>82, Gumi-ro 173 Beon-gil, Bundang-gu</addr-line><addr-line>Seongnam-Si</addr-line><addr-line>Gyeonggi-do</addr-line><country>Republic of Korea</country></aff><aff id="aff2"><institution>Gyeonggi Provincial Medical Center</institution><addr-line>Anseong</addr-line><country>Republic of Korea</country></aff><aff id="aff3"><institution>Gyeonggi Provincial Medical Center</institution><addr-line>Icheon</addr-line><country>Republic of Korea</country></aff><aff id="aff4"><institution>Yonsei University College of Medicine</institution><addr-line>Seoul</addr-line><country>Republic of Korea</country></aff><contrib-group><contrib contrib-type="editor"><name name-style="western"><surname>Mavragani</surname><given-names>Amaryllis</given-names></name></contrib><contrib contrib-type="editor"><name name-style="western"><surname>Law</surname><given-names>Stephanie</given-names></name></contrib></contrib-group><contrib-group><contrib contrib-type="reviewer"><name name-style="western"><surname>Park</surname><given-names>Heejune</given-names></name></contrib></contrib-group><author-notes><corresp>Correspondence to Jee Hyun Suh, MD, PhD, Department of Rehabilitation Medicine, Seoul National University Bundang Hospital, 82, Gumi-ro 173 Beon-gil, Bundang-gu, Seongnam-Si, Gyeonggi-do, Republic of Korea, 82 31-787-7741; <email>jeehyun.suh1@gmail.com</email></corresp></author-notes><pub-date pub-type="collection"><year>2026</year></pub-date><pub-date pub-type="epub"><day>23</day><month>4</month><year>2026</year></pub-date><volume>10</volume><elocation-id>e90110</elocation-id><history><date date-type="received"><day>22</day><month>12</month><year>2025</year></date><date date-type="rev-recd"><day>24</day><month>03</month><year>2026</year></date><date date-type="accepted"><day>25</day><month>03</month><year>2026</year></date></history><copyright-statement>&#x00A9; Jaewon Lee, Sejin Kwon, Jin Young Ko, Yulhyun Park, Jinju Lee, Ju Seok Ryu, Seo Yeon Yoon, Jee Hyun Suh. Originally published in JMIR Formative Research (<ext-link ext-link-type="uri" xlink:href="https://formative.jmir.org">https://formative.jmir.org</ext-link>), 23.4.2026. </copyright-statement><copyright-year>2026</copyright-year><license license-type="open-access" xlink:href="https://creativecommons.org/licenses/by/4.0/"><p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (<ext-link ext-link-type="uri" xlink:href="https://creativecommons.org/licenses/by/4.0/">https://creativecommons.org/licenses/by/4.0/</ext-link>), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR Formative Research, is properly cited. The complete bibliographic information, a link to the original publication on <ext-link ext-link-type="uri" xlink:href="https://formative.jmir.org">https://formative.jmir.org</ext-link>, as well as this copyright and license information must be included.</p></license><self-uri xlink:type="simple" xlink:href="https://formative.jmir.org/2026/1/e90110"/><abstract><sec><title>Background</title><p>Speech sound disorders are common in children and are associated with an increased risk of academic reading difficulties. The COVID-19 pandemic further highlighted the need for remote and digitalized assessment tools. In South Korea, standardized instruments such as the Urimal Test of Articulation and Phonation and Assessment of Phonology and Articulation for children are widely used but have limitations, including reliance on face-to-face evaluation, and the absence of automated scoring.</p></sec><sec><title>Objective</title><p>This study aimed to develop and establish the content validity of an articulation assessment tool that can overcome these limitations and be integrated into digital therapeutics (DTx).</p></sec><sec sec-type="methods"><title>Methods</title><p>A 3-round modified Delphi survey was conducted between July and September 2025 with 92% (23/25) of the invited experts, including 52.2% (12/23) physiatrists and 47.8% (11/23) speech-language pathologists, with a mean professional experience of 10.69 (SD 5.09) years. All participants (23/23, 100%) completed all rounds. Panelists evaluated the appropriateness of word lists, phonological environments, and scoring criteria. Quantitative analyses, including calculations of content validity ratio (CVR), content validity index (CVI), and median and IQR, were performed. Consensus thresholds were set at a CVR of &#x2265;0.39, a CVI of &#x2265;0.78, a median of &#x2265;3.5, and an IQR of &#x2264;1.0. Items were retained only when all 4 criteria were satisfied. While formal qualitative analysis was not performed, the research team internally reviewed and synthesized core keywords and themes from the experts&#x2019; open-ended responses to guide the refinement of items.</p></sec><sec sec-type="results"><title>Results</title><p>These findings were summarized into four key areas: (1) modernization of word stimuli, (2) expansion of phonological coverage, (3) refinement of scoring criteria to reduce ambiguity, and (4) enhancement of result interpretability through visualization. In round 2, a revised 35-word list was evaluated across 25 items, of which 20 (80%) met all consensus criteria. In total, 20% (5/25) of the items failed to meet at least one threshold, including phonological environment adequacy (CVR=0.48; CVI=0.74), scoring redundancy (CVR=0.13; CVI=0.57), usefulness of proportion of whole-word correctness or percentage of word proximity (CVR=0.39; CVI=0.70), contribution of mean phonological length (CVR=0.22; CVI=0.61), and usefulness of feature-based indexes (CVR=0.30; CVI=0.65; IQR 2). Items that reached consensus showed CVR values of 0.57 to 0.91, CVI values of 0.78 to 0.96, a median score of 4, and IQR values of 0 to 1. In round 3, all remaining items achieved consensus.</p></sec><sec sec-type="conclusions"><title>Conclusions</title><p>This Delphi study developed a novel articulation assessment tool with robust content validity. This tool includes updated word stimuli, diverse analysis indexes, and visualization features, thereby enhancing its clinical utility and suitability for integration into artificial intelligence&#x2013;based DTx. By standardizing and digitalizing articulation assessments, this tool has the potential to support personalized and accessible interventions for children with speech sound disorders.</p></sec></abstract><kwd-group><kwd>speech sound disorder</kwd><kwd>Delphi method</kwd><kwd>assessment tool</kwd><kwd>digital therapeutics</kwd><kwd>digital diagnostics</kwd><kwd>digital screening</kwd><kwd>pediatric screening</kwd><kwd>artificial intelligence</kwd><kwd>AI</kwd></kwd-group></article-meta></front><body><sec id="s1" sec-type="intro"><title>Introduction</title><p>Speech sound disorders (SSDs) represent one of the most prevalent communication disorders in childhood, typically manifesting as difficulty in accurate production or inappropriate use of speech sounds relative to developmental expectations for age [<xref ref-type="bibr" rid="ref1">1</xref>]. Recent statistics suggest that approximately 24% of students with disabilities in US public schools are provided with services for speech or language disorders [<xref ref-type="bibr" rid="ref2">2</xref>]. In South Korea, approximately 44.1% of children receiving speech therapy are diagnosed with SSDs, and approximately 9% of children aged 6 years present with articulation difficulties [<xref ref-type="bibr" rid="ref3">3</xref>]. Children with SSDs are at increased risk of developing reading disabilities and other learning difficulties, resulting in significant implications for school readiness and overall academic achievement [<xref ref-type="bibr" rid="ref4">4</xref>,<xref ref-type="bibr" rid="ref5">5</xref>]. Therefore, early identification and intervention for SSDs are vital for reducing the risk of reading disabilities and promoting school readiness.</p><p>The COVID-19 pandemic had a profound global impact in 2020. Environmental factors, including potential effects of mask wearing, have been reported to negatively affect children&#x2019;s cognitive, social, and language development [<xref ref-type="bibr" rid="ref6">6</xref>,<xref ref-type="bibr" rid="ref7">7</xref>]. However, effects related to the COVID-19 pandemic persist, continuing to negatively affect language development in children [<xref ref-type="bibr" rid="ref7">7</xref>]. Telerehabilitation and digital therapeutics (DTx) reduce health care costs, save patients time, and improve accessibility [<xref ref-type="bibr" rid="ref8">8</xref>]. Furthermore, following the COVID-19 pandemic, the use of telerehabilitation increased substantially, with reported rates increasing from 3.4% to 19.3% [<xref ref-type="bibr" rid="ref9">9</xref>]. There is a growing shift toward remote and digital approaches for both assessment and intervention. In this context, appropriate tools for evaluating SSDs are needed to ensure standardized quality, allow for home-based administration, and minimize the need for in-person visits.</p><p>In South Korea, the Urimal Test of Articulation and Phonation (U-TAP) and Assessment of Phonology and Articulation for Children (APAC) are among the most frequently used standardized instruments for the evaluation of SSDs [<xref ref-type="bibr" rid="ref3">3</xref>]. However, despite their widespread use, both tools present certain limitations. They rely heavily on face-to-face administration and perceptual assessment, lack automated or digitalized scoring systems, and have limited applicability in remote or home-based assessment settings. Moreover, normative data are constrained to specific age groups and linguistic contexts, posing challenges for their broader clinical use.</p><p>Beyond these logistical and systemic constraints, these conventional assessments also have technical shortcomings in their analytical frameworks. For instance, both the U-TAP and APAC are time-consuming [<xref ref-type="bibr" rid="ref3">3</xref>], and the U-TAP does not distinguish between medial and final positions when evaluating codas, making it difficult to accurately identify coda errors [<xref ref-type="bibr" rid="ref10">10</xref>]. In contrast, the APAC defines the omission of a medial coda or regressive assimilation of a following medial onset into the coda as typical medial simplification and the omission of a medial onset or progressive assimilation of a medial coda into the following onset as atypical medial simplification [<xref ref-type="bibr" rid="ref10">10</xref>]. On the basis of this framework, the APAC analyzes coda errors separately for medial and final positions. However, it does not differentiate between omission and substitution errors of medial codas as they are all categorized under medial simplification, limiting the detailed analysis of medial coda errors [<xref ref-type="bibr" rid="ref10">10</xref>].</p><p>While recent advances in artificial intelligence (AI), particularly large language models, have expanded the potential for DTx applications in speech-language pathology [<xref ref-type="bibr" rid="ref11">11</xref>], existing research has predominantly focused on intervention efficacy rather than the development of linguistically valid and digitally adaptable assessment frameworks [<xref ref-type="bibr" rid="ref11">11</xref>-<xref ref-type="bibr" rid="ref13">13</xref>]. In particular, recent studies have attempted to apply AI technologies to the detection and assessment of SSDs. In the context of recent advances in digital health, several studies have explored AI-assisted or digital approaches in speech-language pathology. For example, previous work has developed automated SSD detection systems using deep learning models based on automatic speech recognition and audio classification, achieving moderate classification performance (eg, unweighted average recall of approximately 70%) [<xref ref-type="bibr" rid="ref14">14</xref>]. However, these approaches primarily focus on classification tasks (eg, distinguishing children with typical development from those with SSD) and rely on metrics such as proportion of whole-word correctness (PWC) without providing detailed linguistic analysis of articulation error patterns. In addition, their performance is influenced by the accuracy of speech recognition models and predefined thresholds, which may limit interpretability and clinical applicability. Although such studies demonstrate the technical feasibility of AI-based SSD detection, clinically standardized and linguistically grounded assessment tools suitable for widespread diagnostic use remain under development. Therefore, there remains a lack of clinically grounded and digitally adaptable diagnostic frameworks that can accurately capture articulation error patterns while being suitable for digital implementation. In this regard, unlike previous AI-based SSD detection studies focused on classification accuracy, this study aimed to establish a linguistically structured and clinically interpretable assessment framework that can support detailed error pattern analysis and future integration into DTx systems.</p><p>Rapid AI-powered large language models have drawn significant attention for their potential use in DTx [<xref ref-type="bibr" rid="ref11">11</xref>]. To fully realize this potential, it is essential to clarify how such interventions function, what factors affect their effectiveness, and how AI can be systematically integrated into DTx [<xref ref-type="bibr" rid="ref11">11</xref>-<xref ref-type="bibr" rid="ref13">13</xref>]. To develop AI-based DTx for children with SSDs effectively, it is essential to accurately identify each child&#x2019;s specific error patterns and design algorithms to determine appropriate intervention strategies. The most fundamental step in this process is the precise evaluation of a child&#x2019;s articulation error patterns.</p><p>A previous study reported that, while speech-language pathologists demonstrated high intrarater reliability in the evaluation of speech intelligibility, interrater reliability was only moderate to excellent, indicating variability across clinicians [<xref ref-type="bibr" rid="ref15">15</xref>]. Thus, conventional speech-language pathologist&#x2013;based language assessments are limited by variability in scoring and interpretation across clinicians. This limitation may contribute to inequitable and potentially discriminatory access to assessments and treatments across rural and urban regions.</p><p>Taken together, these limitations highlight the need for a novel articulation assessment tool that is not only clinically and linguistically robust but also specifically designed for digital and AI-based applications.</p><p>This study aimed to develop a new articulation development evaluation tool to overcome the technical and clinical limitations of the U-TAP and APAC, facilitate its transition to DTx, and verify its content validity and applicability through a Delphi study.</p></sec><sec id="s2" sec-type="methods"><title>Methods</title><sec id="s2-1"><title>Study Design</title><p>The Delphi survey was conducted in 3 iterative rounds between July 2025 and September 2025. The number of rounds was predefined to a maximum of 3 to ensure a structured consensus-building process. We used a Delphi methodology to systematically gather expert consensus on digital assessment items for articulation and phonological disorders [<xref ref-type="bibr" rid="ref16">16</xref>]. The Delphi technique collects expert consensus anonymously through structured questionnaires and iterative feedback [<xref ref-type="bibr" rid="ref17">17</xref>]. It provides a systematic and objective approach for evaluating content that requires professional judgment and is widely recognized as a standard method for establishing the content validity of clinical assessment tools.</p><p>The initial questionnaire was developed by a multidisciplinary committee consisting of 3 physiatrists and 1 professor of speech-language pathology. The committee drafted the word list based on clinical experience, existing instruments (eg, U-TAP and APAC), and literature on word familiarity. Specifically, candidate words were selected to reflect age-appropriate vocabulary, phonetic balance across consonant positions, and clinical relevance for identifying common articulation error patterns. Relevant literature and existing standardized assessment tools were reviewed to ensure linguistic validity and familiarity for Korean-speaking children.</p><p>To minimize potential psychological biases, such as the bandwagon effect and interference effects; ensure anonymity; and enhance validity, we implemented a web-based modified Delphi process conducted via an online survey platform (Google Forms). All responses were collected anonymously, and participants were blinded to the identities of other panel members. Open- and closed-ended questions were used to facilitate consensus. In round 1, open- and closed-ended questions were used to evaluate the appropriateness of a word list and stimulus composition, diversity and usefulness of phonological environments, and validity of both scoring criteria. Round 2 served as an internal validation stage in which the results of round 1 were analyzed and restructured into closed-ended questions. Summarized feedback from round 1 was provided to participants in round 2 to support iterative consensus building. In round 3, additional feedback was incorporated to refine the items, and a final set was established. In rounds 2 and 3, a 5-point Likert scale was used supplemented with open-ended sections to allow participants to revise, add, or withdraw their opinions.</p><p>Between rounds, both quantitative and qualitative syntheses were performed. The feedback, including statistical summaries and qualitative syntheses, was provided uniformly to all participants to support consensus building. Although a formal qualitative analysis of open-ended responses was not performed, the research team internally reviewed and synthesized core keywords and themes to guide item refinement and establish priorities based on importance and appropriateness. Furthermore, even for items that reached statistical consensus, additional exploratory questions were included in subsequent rounds if the research team identified clinically significant issues requiring further expert deliberation. In addition, panelists were encouraged to consider the scalability and applicability of the assessment framework for future digital and AI-based implementation, including the use of structured, quantifiable scoring systems and standardized output formats that could support automated analysis.</p></sec><sec id="s2-2"><title>Participants</title><p>An expert panel was convened, with specialists identified through the official membership lists of the Korean Academy of Rehabilitation Medicine and the Korean Association of Speech-Language Pathologists. Initially, the research team considered including DTx developers; however, the scope was purposefully narrowed to specialists in rehabilitation medicine and speech-language pathology to focus on establishing the foundational clinical validity of the assessment tool. The panelists were chosen based on their expertise in articulation and phonological disorders and their experience in digital assessment or related fields.</p><p>According to previous studies, homogeneous panels typically require 15 to 30 participants, whereas heterogeneous panels may be sufficiently represented with 5 to 10 participants [<xref ref-type="bibr" rid="ref18">18</xref>]. Given that our Delphi panel included 2 distinct professional groups&#x2014;speech-language pathologists and physiatrists&#x2014;the panel could be considered heterogeneous. Therefore, the inclusion of 23 experts in this study exceeded the recommended size for heterogeneous panels, ensuring both adequacy and diversity of perspectives.</p><p>The criteria for panel selection were as follows: (1) possession of a degree in the relevant specialty, (2) a minimum of 5 years of clinical or research experience, and (3) understanding of study objectives and methodology with consent to participate. Exclusion criteria were as follows: (1) less than 5 years of clinical or research experience, (2) willingness to participate without sufficient understanding of the study objectives or Delphi method, and (3) inability to continue participation due to withdrawal or incomplete responses during the survey rounds. These inclusion and exclusion criteria were established a priori. A total of 25 potential experts were initially invited via email. Of these 25 experts, 23 (92%) agreed to participate, whereas 2 (8%) declined due to scheduling conflicts. No participants were excluded after enrollment, and all 23 experts completed all Delphi rounds and were included in the final analysis.</p></sec><sec id="s2-3"><title>Ethical Considerations</title><p>This study was approved by the Institutional Review Board of Seoul National University Bundang Hospital (B-2507-987-301). All the participating experts were provided with information regarding the purpose of the study, expected time commitment, eligibility criteria, and importance of completing all rounds of the survey. Participation in the Delphi process was voluntary, and informed consent was obtained prior to involvement. Participants received monetary compensation of US $100 for their time and participation in the study. The compensation was determined to be appropriate and not coercive. To ensure anonymity, the survey was conducted online, and each expert was assigned a unique identification code. This procedure allowed responses to remain confidential while maintaining the integrity of the iterative Delphi process. All the responses were collected anonymously to protect participants&#x2019; confidentiality.</p></sec><sec id="s2-4"><title>Data Analysis</title><p>All responses were analyzed using equal weights. Predefined criteria for consensus were established prior to data collection to guide interpretation and agreement thresholds. To ensure content validity, basic descriptive statistics (mean, SD, and median), content validity ratio (CVR), and consensus indexes were examined and analyzed. CVR was calculated to assess the essentiality of each word item based on the number of panelists rating an item as &#x201C;essential.&#x201D; Following the Lawshe formula, the minimum acceptable CVR value for 23 panelists was set at 0.37 [<xref ref-type="bibr" rid="ref19">19</xref>]. Additionally, the content validity index (CVI) was calculated from a percentage of panelists giving each item a rating of 3 or 4 on a 5-point Likert scale, with a CVI of 0.78 or higher indicating high content validity [<xref ref-type="bibr" rid="ref20">20</xref>].</p><p>Consensus levels were evaluated using the median and IQR. A median score of 3.5 or higher was regarded as positive consensus, whereas an IQR of 1.0 or lower indicated a high level of agreement among panelists [<xref ref-type="bibr" rid="ref21">21</xref>].</p><p>Items were retained if they met all 4 criteria: CVR of 0.39 or higher, CVI of 0.78 or higher, median of 3.5 or higher, and IQR of 1.0 or lower. Items that met CVR and CVI thresholds but failed to reach consensus (median or IQR) were revised and re-evaluated in a subsequent round, whereas items that did not meet the CVR or CVI thresholds were either modified or eliminated.</p><p>Although a formal qualitative analysis was not conducted, open-ended responses were reviewed using a structured descriptive approach. Two investigators independently examined the responses to identify recurring themes and clinically relevant insights. Discrepancies were resolved through discussion, and the synthesized findings were used to refine questionnaire items and inform subsequent Delphi rounds.</p></sec></sec><sec id="s3" sec-type="results"><title>Results</title><sec id="s3-1"><title>Overview</title><p>This study aimed to develop a diagnostic tool for children with SSD to support the creation of an AI-based DTx instrument. The three main objectives were to (1) construct a word list and diagnostic tool for children with SSD, (2) evaluate the content validity of the developed tool, and (3) establish diagnostic parameters for differentiating SSD using a Delphi survey process. The findings of the 3 study rounds (<xref ref-type="fig" rid="figure1">Figure 1</xref> shows a flowchart of participation) are described in the following sections.</p><fig position="float" id="figure1"><label>Figure 1.</label><caption><p>Flowchart of the 3-round Delphi process used to develop the articulation assessment tool. Of the 25 recruited experts, 23 (92%) participated in all rounds. The process included an initial assessment (round 1), feedback and revision (round 2), and final consensus (round 3), resulting in the final tool.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="formative_v10i1e90110_fig01.png"/></fig></sec><sec id="s3-2"><title>Round 1</title><p>The first round of the Delphi survey was completed by all 23 experts who initially agreed to participate, achieving a 100% (23/23) response rate for this stage. The panel was asked open-ended and multiple-choice (5-point scale) questions regarding the appropriateness of the 26 initial words, diversity of phonological environments, and validity of the scoring criteria. On the basis of the comprehensive feedback from the 23 experts, the panel emphasized the need to include diphthong words, exclude words with potential dialectal variations (eg, <italic>jajangmyeon</italic> [black bean noodles]), and remove words unfamiliar to modern urban children (eg, <italic>gulttuk</italic> [chimney] and <italic>radio</italic>). They also highlighted the necessity of expanding and supplementing words containing fricative and affricate codes as well as increasing the overall number of test words.</p><p>Furthermore, the experts recommended incorporating listeners&#x2019; subjective ratings of speech clarity, clarifying the scoring criteria for ambiguous productions, improving the scoring system, and presenting the results using graphical visualization. These qualitative insights, synthesized from the open-ended responses, provided the foundational rationale for the modifications implemented in the subsequent round. Notably, several of these refinements, such as the development of clearer scoring criteria and graphical visualization of results, were aligned with the need to generate structured and interpretable data that may be suitable for future AI-based analysis.</p></sec><sec id="s3-3"><title>Round 2</title><p>Following the synthesis of results from the 23 panel members who responded to the round 1 Delphi survey, words from an initial 26-item list that were deemed inappropriate for modern urban children or had the potential to be dialectal were excluded (<xref ref-type="table" rid="table1">Table 1</xref>).</p><table-wrap id="t1" position="float"><label>Table 1.</label><caption><p>Word list for articulation assessment developed through a 3-round Delphi survey among experts in pediatric rehabilitation and speech-language pathology (n=35 words).</p></caption><table id="table1" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom">Target word in Korean</td><td align="left" valign="bottom">Meaning</td><td align="left" valign="bottom">IPA<sup><xref ref-type="table-fn" rid="table1fn1">a</xref></sup> transcription</td></tr></thead><tbody><tr><td align="left" valign="top">&#xD0C0;&#xC870;</td><td align="left" valign="top">Ostrich</td><td align="left" valign="top">[tʰa&#x0320;.d&#x0361;ʑo&#x031E;]</td></tr><tr><td align="left" valign="top">&#xB0C4;&#xBE44;</td><td align="left" valign="top">Pot</td><td align="left" valign="top">[nɛm.bi]</td></tr><tr><td align="left" valign="top">&#xB208;&#xC379;</td><td align="left" valign="top">Eyebrow</td><td align="left" valign="top">[nun.s&#x0348;ʌp&#x031A;]</td></tr><tr><td align="left" valign="top">&#xB538;&#xAE30;</td><td align="left" valign="top">Strawberry</td><td align="left" valign="top">[t&#x0348;a&#x0320;l.ɡi]</td></tr><tr><td align="left" valign="top">&#xAD6D;&#xC790;</td><td align="left" valign="top">Ladle</td><td align="left" valign="top">[kuk&#x031A;.t&#x0361;ɕ&#x0348;a&#x0320;]</td></tr><tr><td align="left" valign="top">&#xB2E4;&#xB9AC;</td><td align="left" valign="top">Leg and bridge</td><td align="left" valign="top">[ta&#x0320;.ɾi]</td></tr><tr><td align="left" valign="top">&#xC790;&#xB3D9;&#xCC28;</td><td align="left" valign="top">Car</td><td align="left" valign="top">[t&#x0361;ɕa&#x0320;.do&#x031E;&#x014B;.t&#x0361;ɕʰa&#x0320;]</td></tr><tr><td align="left" valign="top">&#xC528;&#xC557;</td><td align="left" valign="top">Seed</td><td align="left" valign="top">[ɕ&#x0348;i.a&#x0320;t&#x031A;]</td></tr><tr><td align="left" valign="top">&#xBC14;&#xB098;&#xB098;</td><td align="left" valign="top">Banana</td><td align="left" valign="top">[pa&#x0320;.na&#x0320;.na&#x0320;]</td></tr><tr><td align="left" valign="top">&#xB808;&#xBAAC;</td><td align="left" valign="top">Lemon</td><td align="left" valign="top">[ɾe&#x031E;.mo&#x031E;n]</td></tr><tr><td align="left" valign="top">&#xCF54;&#xB07C;&#xB9AC;</td><td align="left" valign="top">Elephant</td><td align="left" valign="top">[kʰo&#x031E;.k&#x0348;i.ɾi]</td></tr><tr><td align="left" valign="top">&#xACF0;</td><td align="left" valign="top">Bear</td><td align="left" valign="top">[ko&#x031E;m]</td></tr><tr><td align="left" valign="top">&#xC0AC;&#xD0D5;</td><td align="left" valign="top">Candy</td><td align="left" valign="top">[sa&#x0320;.tʰa&#x0320;&#x014B;]</td></tr><tr><td align="left" valign="top">&#xBC84;&#xC12F;</td><td align="left" valign="top">Mushroom</td><td align="left" valign="top">[pʌ.sʌt&#x031A;]</td></tr><tr><td align="left" valign="top">&#xD3AD;&#xADC4;</td><td align="left" valign="top">Penguin</td><td align="left" valign="top">[pʰe&#x031E;&#x014B;.ɡɥin]</td></tr><tr><td align="left" valign="top">&#xB9C8;&#xC774;&#xD06C;</td><td align="left" valign="top">Microphone</td><td align="left" valign="top">[ma&#x0320;.i.kʰɯ]</td></tr><tr><td align="left" valign="top">&#xC5F0;&#xD544;</td><td align="left" valign="top">Pencil</td><td align="left" valign="top">[jʌn.pʰil]</td></tr><tr><td align="left" valign="top">&#xAE4C;&#xCE58;</td><td align="left" valign="top">Magpie</td><td align="left" valign="top">[k&#x0348;a&#x0320;.t&#x0361;ɕʰi]</td></tr><tr><td align="left" valign="top">&#xBE75;</td><td align="left" valign="top">Bread</td><td align="left" valign="top">[p&#x0348;a&#x0320;&#x014B;]</td></tr><tr><td align="left" valign="top">&#xC544;&#xBE60;</td><td align="left" valign="top">Dad</td><td align="left" valign="top">[a&#x0320;p&#x0348;a&#x0320;]</td></tr><tr><td align="left" valign="top">&#xCC45;&#xC0C1;</td><td align="left" valign="top">Desk</td><td align="left" valign="top">[t&#x0361;ɕʰɛk&#x031A;.s&#x0348;a&#x0320;&#x014B;]</td></tr><tr><td align="left" valign="top">&#xCC0C;&#xAC1C;</td><td align="left" valign="top">Stew</td><td align="left" valign="top">[t&#x0361;ɕ&#x0348;i.ɡɛ]</td></tr><tr><td align="left" valign="top">&#xC624;&#xB69D;&#xC774;</td><td align="left" valign="top">Roly-poly toy</td><td align="left" valign="top">[o&#x031E;.t&#x0348;u.ɡi]</td></tr><tr><td align="left" valign="top">&#xD558;&#xB9C8;</td><td align="left" valign="top">Hippopotamus</td><td align="left" valign="top">[ha&#x0320;.ma&#x0320;]</td></tr><tr><td align="left" valign="top">&#xAD6C;&#xAE09;&#xCC28;</td><td align="left" valign="top">Ambulance</td><td align="left" valign="top">[ku.ɡɯp&#x031A;.t&#x0361;ɕʰa&#x0320;]</td></tr><tr><td align="left" valign="top">&#xC218;&#xBC15;</td><td align="left" valign="top">Watermelon</td><td align="left" valign="top">[su.ba&#x0320;k&#x031A;]</td></tr><tr><td align="left" valign="top">&#xACFC;&#xC790;</td><td align="left" valign="top">Snack</td><td align="left" valign="top">[kwa&#x0320;.d&#x0361;ʑa&#x0320;]</td></tr><tr><td align="left" valign="top">&#xB3FC;&#xC9C0;</td><td align="left" valign="top">Pig</td><td align="left" valign="top">[twe&#x031E;.d&#x0361;ʑi]</td></tr><tr><td align="left" valign="top">&#xC694;&#xAC70;&#xD2B8;</td><td align="left" valign="top">Yogurt</td><td align="left" valign="top">[jo&#x031E;.ɡʌ.tʰɯ]</td></tr><tr><td align="left" valign="top">&#xADC0;</td><td align="left" valign="top">Ear</td><td align="left" valign="top">[kɥi]</td></tr><tr><td align="left" valign="top">&#xC5EC;&#xC6B0;</td><td align="left" valign="top">Fox</td><td align="left" valign="top">[jʌ.u]</td></tr><tr><td align="left" valign="top">&#xC6B0;&#xC720;</td><td align="left" valign="top">Milk</td><td align="char" char="." valign="top">[u.ju]</td></tr><tr><td align="left" valign="top">&#xC57C;&#xAD6C;</td><td align="left" valign="top">Baseball</td><td align="left" valign="top">[ja&#x0320;.ɡu]</td></tr><tr><td align="left" valign="top">&#xC6D0;&#xC22D;&#xC774;</td><td align="left" valign="top">Monkey</td><td align="left" valign="top">[wʌn.su.&#x014B;i]</td></tr><tr><td align="left" valign="top">&#xC758;&#xC790;</td><td align="left" valign="top">Chair</td><td align="left" valign="top">[ɰi.d&#x0361;ʑa&#x0320;]</td></tr></tbody></table><table-wrap-foot><fn id="table1fn1"><p><sup>a</sup>IPA: International Phonetic Alphabet.</p></fn></table-wrap-foot></table-wrap><p>These were replaced with alternative items, along with additional words representing diverse articulatory environments (eg, <italic>remon</italic> [&#xB808;&#xBAAC;], meaning &#x201C;lemon,&#x201D; <italic>ottugi</italic> [&#xC624;&#xB69D;&#xC774;], meaning &#x201C;roly-poly toy,&#x201D; and <italic>jjigae</italic> [&#xCC0C;&#xAC1C;], meaning &#x201C;stew&#x201D;). Furthermore, words containing diphthongs (eg, <italic>yagu</italic> [&#xC57C;&#xAD6C;], meaning &#x201C;baseball,&#x201D; <italic>uija</italic> [&#xC758;&#xC790;], meaning &#x201C;chair,&#x201D; <italic>wonsungi</italic> [&#xC6D0;&#xC22D;&#xC774;], meaning &#x201C;monkey,&#x201D; <italic>yogeoteu</italic> [&#xC694;&#xAC70;&#xD2B8;], meaning &#x201C;yogurt,&#x201D; <italic>dwaeji</italic> [&#xB3FC;&#xC9C0;], meaning &#x201C;pig,&#x201D; and <italic>gwaja</italic> [&#xACFC;&#xC790;], meaning &#x201C;snack&#x201D; or &#x201C;cracker&#x201D;) were incorporated, resulting in a final test list of 35 words. To facilitate a clearer interpretation of speech outcomes by both caregivers and clinicians, visual graphs were integrated into the scoring results to display consonant and vowel accuracies (<xref ref-type="table" rid="table2">Table 2</xref>).</p><table-wrap id="t2" position="float"><label>Table 2.</label><caption><p>Consensus articulation assessment measures and scoring system established through a 3-round Delphi survey.</p></caption><table id="table2" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom">Articulation outcome category and subcategory</td><td align="left" valign="bottom">Denominator</td><td align="left" valign="bottom">Unit used for scoring</td></tr></thead><tbody><tr><td align="left" valign="top" colspan="3">Accuracy</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Consonant accuracy (PCC<sup><xref ref-type="table-fn" rid="table2fn1">a</xref></sup>)</td><td align="left" valign="top">48</td><td align="left" valign="top">Percentage</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Revised consonant accuracy (PCC-revised)</td><td align="left" valign="top">48</td><td align="left" valign="top">Percentage</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Total consonant accuracy (total PCC)</td><td align="left" valign="top">48</td><td align="left" valign="top">Percentage</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Total revised consonant accuracy (total PCC-revised)</td><td align="left" valign="top">48</td><td align="left" valign="top">Percentage</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Vowel accuracy (PVC<sup><xref ref-type="table-fn" rid="table2fn2">b</xref></sup>)</td><td align="left" valign="top">14</td><td align="left" valign="top">Percentage</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Diphthong accuracy</td><td align="left" valign="top">0</td><td align="left" valign="top">Percentage</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Word accuracy (PWC<sup><xref ref-type="table-fn" rid="table2fn3">c</xref></sup>)</td><td align="left" valign="top">26</td><td align="left" valign="top">Percentage</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>PMLU<sup><xref ref-type="table-fn" rid="table2fn4">d</xref></sup></td><td align="left" valign="top">26</td><td align="left" valign="top">Percentage</td></tr><tr><td align="left" valign="top" colspan="3">Articulation manner</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Plosive</td><td align="left" valign="top">23</td><td align="left" valign="top">Percentage</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Fricative</td><td align="left" valign="top">5</td><td align="left" valign="top">Percentage</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Affricate</td><td align="left" valign="top">6</td><td align="left" valign="top">Percentage</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Nasal</td><td align="left" valign="top">10</td><td align="left" valign="top">Percentage</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Liquid</td><td align="left" valign="top">4</td><td align="left" valign="top">Percentage</td></tr><tr><td align="left" valign="top" colspan="3">Articulation place</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Bilabial</td><td align="left" valign="top">12</td><td align="left" valign="top">Percentage</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Alveolar</td><td align="left" valign="top">19</td><td align="left" valign="top">Percentage</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Palatal</td><td align="left" valign="top">6</td><td align="left" valign="top">Percentage</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Velar</td><td align="left" valign="top">10</td><td align="left" valign="top">Percentage</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Glottal</td><td align="left" valign="top">1</td><td align="left" valign="top">Percentage</td></tr><tr><td align="left" valign="top" colspan="3">Phonological features</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Continuancy</td><td align="left" valign="top">6</td><td align="left" valign="top">Percentage</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Delayed release</td><td align="left" valign="top">6</td><td align="left" valign="top">Percentage</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Laterality</td><td align="left" valign="top">3</td><td align="left" valign="top">Percentage</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Coronal</td><td align="left" valign="top">23</td><td align="left" valign="top">Percentage</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Anterior</td><td align="left" valign="top">27</td><td align="left" valign="top">Percentage</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Tenseness</td><td align="left" valign="top">18</td><td align="left" valign="top">Percentage</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Aspiration</td><td align="left" valign="top">8</td><td align="left" valign="top">Percentage</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>High tongue</td><td align="left" valign="top">14</td><td align="left" valign="top">Percentage</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Back tongue</td><td align="left" valign="top">10</td><td align="left" valign="top">Percentage</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Roundedness</td><td align="left" valign="top">3</td><td align="left" valign="top">Percentage</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>High</td><td align="left" valign="top">6</td><td align="left" valign="top">Percentage</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Low</td><td align="left" valign="top">1</td><td align="left" valign="top">Percentage</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Back</td><td align="left" valign="top">7</td><td align="left" valign="top">Percentage</td></tr><tr><td align="left" valign="top" colspan="3">Error patterns&#x2014;substitution</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Plosivization</td><td align="left" valign="top">21</td><td align="left" valign="top">Percentage</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Fricativization</td><td align="left" valign="top">37</td><td align="left" valign="top">Percentage</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Affricativization</td><td align="left" valign="top">37</td><td align="left" valign="top">Percentage</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Nasalization</td><td align="left" valign="top">36</td><td align="left" valign="top">Percentage</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Liquidization</td><td align="left" valign="top">40</td><td align="left" valign="top">Percentage</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Bilabialization</td><td align="left" valign="top">33</td><td align="left" valign="top">Percentage</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Alveolarization</td><td align="left" valign="top">26</td><td align="left" valign="top">Percentage</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Palatalization</td><td align="left" valign="top">37</td><td align="left" valign="top">Percentage</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Velarization</td><td align="left" valign="top">35</td><td align="left" valign="top">Percentage</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Glottalization</td><td align="left" valign="top">41</td><td align="left" valign="top">Percentage</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Tense consonantization</td><td align="left" valign="top">34</td><td align="left" valign="top">Percentage</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Aspirated consonantization</td><td align="left" valign="top">33</td><td align="left" valign="top">Percentage</td></tr><tr><td align="left" valign="top" colspan="3">Error patterns&#x2014;assimilation</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Plosive assimilation</td><td align="left" valign="top">16</td><td align="left" valign="top">Percentage</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Fricative assimilation</td><td align="left" valign="top">5</td><td align="left" valign="top">Percentage</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Affricate assimilation</td><td align="left" valign="top">8</td><td align="left" valign="top">Percentage</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Nasal assimilation</td><td align="left" valign="top">12</td><td align="left" valign="top">Percentage</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Liquid assimilation</td><td align="left" valign="top">9</td><td align="left" valign="top">Percentage</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Bilabial assimilation</td><td align="left" valign="top">11</td><td align="left" valign="top">Percentage</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Alveolar assimilation</td><td align="left" valign="top">13</td><td align="left" valign="top">Percentage</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Palatal assimilation</td><td align="left" valign="top">8</td><td align="left" valign="top">Percentage</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Velar assimilation</td><td align="left" valign="top">14</td><td align="left" valign="top">Percentage</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Glottal assimilation</td><td align="left" valign="top">2</td><td align="left" valign="top">Percentage</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Tense assimilation</td><td align="left" valign="top">9</td><td align="left" valign="top">Percentage</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Aspirated assimilation</td><td align="left" valign="top">9</td><td align="left" valign="top">Percentage</td></tr></tbody></table><table-wrap-foot><fn id="table2fn1"><p><sup>a</sup>PCC: percentage of consonants correct.</p></fn><fn id="table2fn2"><p><sup>b</sup>PVC: percentage of vowels correct.</p></fn><fn id="table2fn3"><p><sup>c</sup>PWC: proportion of whole-word correctness.</p></fn><fn id="table2fn4"><p><sup>d</sup>PMLU: phonological mean length of utterance.</p></fn></table-wrap-foot></table-wrap><p>Subsequently, the round 2 Delphi survey used a 5-point Likert scale. This survey was conducted with the same 23 experts who completed the previous round, maintaining a 100% (23/23) retention rate for this round. The survey comprised 7 items that evaluated the appropriateness of the word list and stimulus composition, 5 items assessing the diversity and adequacy of phonological environments within the word list, and 13 items examining the validity of the scoring criteria (Table S2 in <xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref>).</p><p>In the second-round Delphi survey, most items demonstrated strong expert agreement, meeting the predetermined consensus thresholds of CVR of 0.39 or higher, CVI of 0.78 or higher, median of 3.5 or higher, and IQR of 1.0 or lower. These results indicate that most items of the word list, phonological environments, and scoring criteria were assessed to be appropriate and valid for clinical applications.</p><p>However, no consensus was reached regarding 5 items. Specifically, experts did not agree on whether (1) the phonological environments adequately reflected common child speech errors, (2) the scoring rubric was excessively detailed or redundant, (3) analyses using PWC or percentage of word proximity were useful for setting treatment goals, (4) the mean phonological length contributed effectively to the evaluation of phonological complexity, and (5) feature-based indexes provided practical benefits for error pattern analysis (<xref ref-type="table" rid="table3">Table 3</xref>). The specific results for each round, including items that achieved consensus and those that did not, are presented in Table 3 to provide a transparent overview of the refinement process.</p><table-wrap id="t3" position="float"><label>Table 3.</label><caption><p>Outcomes of the second-round Delphi survey evaluating the appropriateness of the word list, phonological environments, and scoring criteria for articulation assessment.</p></caption><table id="table3" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom">Category and survey question</td><td align="left" valign="bottom">CVR<sup><xref ref-type="table-fn" rid="table3fn1">a</xref></sup></td><td align="left" valign="bottom">CVI<sup><xref ref-type="table-fn" rid="table3fn2">b</xref></sup></td><td align="left" valign="bottom">Median (IQR)</td><td align="left" valign="bottom">Agreement (proportion, %)</td></tr></thead><tbody><tr><td align="left" valign="top" colspan="5">Word list and stimulus appropriateness</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Word list matches assessment purpose</td><td align="left" valign="top">0.83</td><td align="left" valign="top">0.91</td><td align="left" valign="top">4 (0)</td><td align="left" valign="top">1</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Word list appropriate for the ages of 2&#x2010;18 y; familiar vocabulary</td><td align="left" valign="top">0.65</td><td align="left" valign="top">0.83</td><td align="left" valign="top">4 (0)</td><td align="left" valign="top">1</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Words difficult to visualize or confusing</td><td align="left" valign="top">0.83</td><td align="left" valign="top">0.91</td><td align="left" valign="top">4 (0)</td><td align="left" valign="top">1</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Presentation order appropriate for flow and difficulty</td><td align="left" valign="top">0.83</td><td align="left" valign="top">0.91</td><td align="left" valign="top">4 (0)</td><td align="left" valign="top">1</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Word list appropriate for frequency and exposure by age</td><td align="left" valign="top">0.91</td><td align="left" valign="top">0.96</td><td align="left" valign="top">4 (1)</td><td align="left" valign="top">0.75</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Word length appropriate</td><td align="left" valign="top">0.83</td><td align="left" valign="top">0.91</td><td align="left" valign="top">4 (1)</td><td align="left" valign="top">0.75</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Factors to prioritize for engaging stimulus words</td><td align="left" valign="top">0.83</td><td align="left" valign="top">0.91</td><td align="left" valign="top">4 (0.5)</td><td align="left" valign="top">0.88</td></tr><tr><td align="left" valign="top" colspan="5">Phonological environments</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Adequate coverage of consonant positions, places, manners, and features</td><td align="left" valign="top">0.83</td><td align="left" valign="top">0.91</td><td align="left" valign="top">4 (0)</td><td align="left" valign="top">1</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Selection of target vowels valid</td><td align="left" valign="top">0.74</td><td align="left" valign="top">0.87</td><td align="left" valign="top">4 (1)</td><td align="left" valign="top">0.75</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Environments reflect common child speech errors</td><td align="left" valign="top">0.48</td><td align="left" valign="top">0.74<sup><xref ref-type="table-fn" rid="table3fn3">c</xref></sup></td><td align="left" valign="top">4 (0.5)</td><td align="left" valign="top">0.88</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Exclude dialectal or regional words</td><td align="left" valign="top">0.74</td><td align="left" valign="top">0.87</td><td align="left" valign="top">4 (0)</td><td align="left" valign="top">1</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Option for diphthongs due to increased task difficulty or vocabulary load</td><td align="left" valign="top">0.74</td><td align="left" valign="top">0.87</td><td align="left" valign="top">4 (1)</td><td align="left" valign="top">0.8</td></tr><tr><td align="left" valign="top" colspan="5">Scoring criteria</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Understanding of indexes (PCC<sup><xref ref-type="table-fn" rid="table3fn4">d</xref></sup>, PMLU<sup><xref ref-type="table-fn" rid="table3fn5">e</xref></sup>, PWP<sup><xref ref-type="table-fn" rid="table3fn6">f</xref></sup>, and PVC<sup><xref ref-type="table-fn" rid="table3fn7">g</xref></sup>)</td><td align="left" valign="top">0.65</td><td align="left" valign="top">0.83</td><td align="left" valign="top">4 (0)</td><td align="left" valign="top">1</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Usefulness of indexes in clinical practice</td><td align="left" valign="top">0.65</td><td align="left" valign="top">0.83</td><td align="left" valign="top">4 (1)</td><td align="left" valign="top">0.75</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Validity of consonant error classification</td><td align="left" valign="top">0.91</td><td align="left" valign="top">0.96</td><td align="left" valign="top">4 (0)</td><td align="left" valign="top">1</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Clarity of scoring criteria (place, manner, phonation, and features)</td><td align="left" valign="top">0.91</td><td align="left" valign="top">0.96</td><td align="left" valign="top">4 (1)</td><td align="left" valign="top">0.75</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Appropriateness of scoring method</td><td align="left" valign="top">0.57</td><td align="left" valign="top">0.78</td><td align="left" valign="top">4 (0)</td><td align="left" valign="top">1</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Sufficiency of indexes for clinical interpretation</td><td align="left" valign="top">0.57</td><td align="left" valign="top">0.78</td><td align="left" valign="top">4 (0.5)</td><td align="left" valign="top">0.88</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Scoring rubric excessively detailed or redundant</td><td align="left" valign="top">0.13<sup><xref ref-type="table-fn" rid="table3fn3">c</xref></sup></td><td align="left" valign="top">0.57<sup><xref ref-type="table-fn" rid="table3fn3">c</xref></sup></td><td align="left" valign="top">4 (1)</td><td align="left" valign="top">0.75</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Appropriateness for caregiver explanation</td><td align="left" valign="top">0.65</td><td align="left" valign="top">0.83</td><td align="left" valign="top">4 (1)</td><td align="left" valign="top">0.75</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Usefulness of PWC<sup><xref ref-type="table-fn" rid="table3fn8">h</xref></sup> and PWP for goal setting</td><td align="left" valign="top">0.39</td><td align="left" valign="top">0.70<sup><xref ref-type="table-fn" rid="table3fn3">c</xref></sup></td><td align="left" valign="top">4 (1)</td><td align="left" valign="top">0.75</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Contribution of mean phonological length to complexity</td><td align="left" valign="top">0.22<sup><xref ref-type="table-fn" rid="table3fn3">c</xref></sup></td><td align="left" valign="top">0.61<sup><xref ref-type="table-fn" rid="table3fn3">c</xref></sup></td><td align="left" valign="top">4 (1)</td><td align="left" valign="top">0.75</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Usefulness of accuracy by place and manner</td><td align="left" valign="top">0.91</td><td align="left" valign="top">0.96</td><td align="left" valign="top">4 (1)</td><td align="left" valign="top">0.75</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Contribution of feature-based indexes to error analysis</td><td align="left" valign="top">0.30<sup><xref ref-type="table-fn" rid="table3fn3">c</xref></sup></td><td align="left" valign="top">0.65<sup><xref ref-type="table-fn" rid="table3fn3">c</xref></sup></td><td align="left" valign="top">4 (2)<sup><xref ref-type="table-fn" rid="table3fn3">c</xref></sup></td><td align="left" valign="top">0.5<sup><xref ref-type="table-fn" rid="table3fn3">c</xref></sup></td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>Need for comprehensive report format</td><td align="left" valign="top">0.57</td><td align="left" valign="top">0.78</td><td align="left" valign="top">4 (1)</td><td align="left" valign="top">0.75</td></tr></tbody></table><table-wrap-foot><fn id="table3fn1"><p><sup>a</sup>CVR: content validity ratio.</p></fn><fn id="table3fn2"><p><sup>b</sup>CVI: content validity index.</p></fn><fn id="table3fn3"><p><sup>c</sup>Did not meet the predefined validity criteria.</p></fn><fn id="table3fn4"><p><sup>d</sup>PCC: percentage of consonants correct.</p></fn><fn id="table3fn5"><p><sup>e</sup>PMLU: phonological mean length of utterance.</p></fn><fn id="table3fn6"><p><sup>f</sup>PWP: percentage of word proximity.</p></fn><fn id="table3fn7"><p><sup>g</sup>PVC: percentage of vowels correct.</p></fn><fn id="table3fn8"><p><sup>h</sup>PWC: proportion of whole-word correctness.</p></fn></table-wrap-foot></table-wrap></sec><sec id="s3-4"><title>Round 3</title><p>The third-round Delphi survey was conducted with the same 23 experts, maintaining a 100% (23/23) retention rate across all 3 rounds of the study. This final round was developed based on the results of the second round, incorporating modifications that were refined through the internal synthesis of expert keywords and thematic feedback from previous open-ended responses. Following this process, percentage of word proximity was removed, and the measurement of mean phonological length was modified to a supplementary index.</p><p>The five items addressed in the third-round Delphi survey were as follows: (1) word-level phonological environments adequately reflect contexts in which children&#x2019;s speech errors commonly occur, (2) the scoring rubric is excessively detailed or redundant, (3) analyses using PWC are useful for establishing treatment goals, (4) the measurement of mean phonological length can serve as a supplementary indicator, and (5) consonant feature analysis contributes to the identification of error patterns. As a result, all 5 items reached consensus, meeting the thresholds of CVR of 0.39 or higher, CVI of 0.78 or higher, median of 3.5 or higher, and IQR of 1.0 or lower. This achieved the final expert consensus required for the development of the assessment tool.</p></sec></sec><sec id="s4" sec-type="discussion"><title>Discussion</title><sec id="s4-1"><title>Principal Findings</title><p>In this study, we developed and validated stimulus words and an interpretation framework for articulation assessment in children with SSD using a 3-round Delphi process [<xref ref-type="bibr" rid="ref22">22</xref>]. The resulting tool comprises 35 words that achieved expert consensus for developmental appropriateness and comprehensibility and incorporates updated stimuli, diverse analytical indexes, and visualization features to enhance clinical applicability and potential integration into AI-based DTx.</p><p>Compared with conventional standardized assessments such as the U-TAP and APAC, this tool introduces several important methodological and practical differences. First, while existing tools primarily rely on face-to-face administration and perceptual evaluation, our tool was designed with a structured and digitally adaptable framework, facilitating future integration into AI-based systems. Second, unlike the U-TAP, which does not distinguish between medial and final coda positions, this framework allows for more detailed positional analysis of phonological errors [<xref ref-type="bibr" rid="ref10">10</xref>]. Third, although the APAC distinguishes positional coda errors, it does not differentiate between omission and substitution errors in medial codas; in contrast, our framework enables more granular classification of error patterns through structured indexes and feature-based analysis [<xref ref-type="bibr" rid="ref10">10</xref>]. Furthermore, it enables the generation of diverse articulation analysis outcomes that can serve as a foundation for AI-based DTx. By standardizing and digitalizing articulation assessments, this tool has the potential to support personalized and accessible interventions for children with SSDs. Importantly, the structured and multidimensional output generated by this framework may facilitate its use as input data for future AI-based diagnostic or decision support systems. These findings extend prior AI-based SSD research by moving beyond classification-focused approaches toward a clinically interpretable and linguistically structured assessment framework. While previous studies have demonstrated the feasibility of automated detection, they have been limited in capturing detailed articulation error patterns and supporting clinical decision-making [<xref ref-type="bibr" rid="ref14">14</xref>]. In contrast, this study emphasizes structured error analysis and interpretability, which are essential for clinical applicability and future integration into DTx systems.</p><p>These findings should be interpreted in relation to the gaps identified in the Introduction section. Specifically, this study addresses the lack of standardized and digitally adaptable assessment frameworks by establishing a structured scoring system and expanding phonological coverage. In addition, the incorporation of visualization features and multiple analytical indexes may contribute to improved interpretability and potential reduction in ambiguity in clinical assessment. However, some previously identified limitations remain unresolved [<xref ref-type="bibr" rid="ref15">15</xref>]. In particular, interrater variability in perceptual judgment was not directly addressed, and automated or AI-based scoring systems were not implemented or validated in this study. Therefore, this work should be considered a foundational step toward digital assessment rather than a fully operational digital diagnostic system. In particular, while this study establishes a structured and AI-compatible assessment framework, further work is required to implement and validate automated or AI-driven diagnostic models based on this framework.</p><p>Compared with conventional assessment tools, which are often limited to static and paper-based formats, this tool provides a flexible and extensible digital framework. This aligns with recent trends in digital health and AI-assisted diagnostics, where standardized input and scalable data structures are essential for model training and clinical deployment.</p><p>This study holds significant value in establishing clinical content validity and a systematic derivation methodology, addressing the limitations of existing analog tools and the challenges of digital transition. To ensure that this tool functions as a reliable diagnostic system within a DTx environment, several directions for future research are proposed. First, the clinical reliability and validity of the word list and assessment system should be empirically verified through correlation analyses using existing standardized instruments (eg, U-TAP and APAC). In addition, large-scale normative data across different ages and genders should be collected to establish percentile-based reference values and objective diagnostic criteria for distinguishing typical from delayed development. Second, optimization studies addressing technical variables in real-world digital environments are required. These include evaluating the effects of environmental factors, such as background noise and device-related variability (eg, microphone performance), on the accuracy of AI-based automated scoring. Furthermore, systematic investigations are needed to establish technical guidelines that ensure consistent and reliable performance in home-based settings. The relationship between the linguistic properties of the target words and speech recognition performance should also be examined to support robust implementation. Finally, future studies should incorporate user interface and user experience design considerations. In particular, the cognitive load imposed on children in screen-based assessment environments should be carefully evaluated. Factors such as visual layout, stimulus presentation timing, and interaction design may influence attention, comprehension, and response accuracy. Accordingly, interdisciplinary collaboration with user interface and user experience specialists and educational technology developers will be essential to optimize usability and real-world applicability in DTx systems.</p><p>In addition, the interpretation of visualized assessment results should be approached with caution. Graphical outputs are intended to support understanding rather than serve as stand-alone diagnostic indicators, and clinical interpretation by qualified professionals remains essential to avoid potential misinterpretation. Furthermore, given that children&#x2019;s voice data constitute sensitive personal information, future implementation of this tool within DTx environments must incorporate robust data security measures and ethical management frameworks. This includes ensuring data privacy, secure storage, and compliance with relevant regulatory standards.</p></sec><sec id="s4-2"><title>Limitations</title><p>This study has several limitations. First, although the Delphi method ensures content validity through expert consensus, it does not provide empirical validation of clinical reliability or diagnostic accuracy. Second, the panel consisted of experts from 2 professional groups, which, while appropriate for the study aims, may limit the generalizability of the findings to other disciplines or international contexts. Third, formal qualitative analysis of open-ended responses was not conducted, which may have limited the depth of thematic interpretation.</p></sec><sec id="s4-3"><title>Conclusions</title><p>Overall, this study provides a foundational framework for the development of digitally enabled articulation assessment tools and highlights the potential for integrating clinically validated instruments into AI-based therapeutic systems.</p><p>Future studies should verify the noninferiority of the newly developed articulation assessment tool by comparing its outcomes, such as consonant and vowel accuracy, with those of standardized instruments (eg, U-TAP). Additionally, the results of evaluations by speech-language pathologists should be compared with those generated by an AI-based assessment tool within the application to further establish its clinical validity. Furthermore, integration of this tool into DTx should be followed by clinical trials to evaluate its real-world effectiveness.</p></sec></sec></body><back><ack><p>The authors confirm that no generative artificial intelligence tools or large language models were used at any stage of the research or manuscript preparation.</p></ack><notes><sec><title>Funding</title><p>This work was supported by a National IT Industry Promotion Agency grant funded by the Korean government (Ministry of Science and ICT; PJT-25-040355). This work was also supported by grant 14-2025-0015 from the Seoul National University Bundang Hospital Research Fund.</p></sec><sec><title>Data Availability</title><p>The datasets generated or analyzed during this study are available from the corresponding author on reasonable request.</p></sec></notes><fn-group><fn fn-type="con"><p>Conceptualization: JYK, YP, Jinju Lee, JSR, SYY, JHS</p><p>Data curation: SK, JYK, YP, Jinju Lee, JSR, SYY</p><p>Formal analysis: SK</p><p>Methodology: JYK, YP, Jinju Lee, JSR, SYY, JHS</p><p>Writing&#x2014;original draft: Jaewon Lee, JHS</p><p>Writing&#x2014;review and editing: JHS</p></fn><fn fn-type="conflict"><p>None declared.</p></fn></fn-group><glossary><title>Abbreviations</title><def-list><def-item><term id="abb1">AI</term><def><p>artificial intelligence</p></def></def-item><def-item><term id="abb2">APAC</term><def><p>Assessment of Phonology and Articulation for Children</p></def></def-item><def-item><term id="abb3">CVI</term><def><p>content validity index</p></def></def-item><def-item><term id="abb4">CVR</term><def><p>content validity ratio</p></def></def-item><def-item><term id="abb5">DELPHISTAR</term><def><p>Delphi Studies in Social and Health Sciences&#x2013;Recommendations for a Standardized Reporting</p></def></def-item><def-item><term id="abb6">DTx</term><def><p>digital therapeutics</p></def></def-item><def-item><term id="abb7">PWC</term><def><p>proportion of whole-word correctness</p></def></def-item><def-item><term id="abb8">SSD</term><def><p>speech sound disorder</p></def></def-item><def-item><term id="abb9">U-TAP</term><def><p>Urimal Test of Articulation and Phonation</p></def></def-item></def-list></glossary><ref-list><title>References</title><ref id="ref1"><label>1</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Waring</surname><given-names>R</given-names> </name><name name-style="western"><surname>Knight</surname><given-names>R</given-names> </name></person-group><article-title>How should children with speech sound disorders be classified? A review and critical evaluation of current classification systems</article-title><source>Int J Lang Commun Disord</source><year>2013</year><month>01</month><volume>48</volume><issue>1</issue><fpage>25</fpage><lpage>40</lpage><pub-id pub-id-type="doi">10.1111/j.1460-6984.2012.00195.x</pub-id><pub-id pub-id-type="medline">23317382</pub-id></nlm-citation></ref><ref id="ref2"><label>2</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Byers</surname><given-names>BA</given-names> </name><name name-style="western"><surname>Bellon-Harn</surname><given-names>ML</given-names> </name><name name-style="western"><surname>Allen</surname><given-names>M</given-names> </name><name name-style="western"><surname>Saar</surname><given-names>KW</given-names> </name><name name-style="western"><surname>Manchaiah</surname><given-names>V</given-names> </name><name name-style="western"><surname>Rodrigo</surname><given-names>H</given-names> </name></person-group><article-title>A comparison of intervention intensity and service delivery models with school-age children with speech sound disorders in a school setting</article-title><source>Lang Speech Hear Serv Sch</source><year>2021</year><month>04</month><day>20</day><volume>52</volume><issue>2</issue><fpage>529</fpage><lpage>541</lpage><pub-id pub-id-type="doi">10.1044/2020_LSHSS-20-00057</pub-id><pub-id pub-id-type="medline">33497584</pub-id></nlm-citation></ref><ref id="ref3"><label>3</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Kim</surname><given-names>DH</given-names> </name><name name-style="western"><surname>Jeong</surname><given-names>JW</given-names> </name><name name-style="western"><surname>Kang</surname><given-names>D</given-names> </name><etal/></person-group><article-title>Usefulness of automatic speech recognition assessment of children with speech sound disorders: validation study</article-title><source>J Med Internet Res</source><year>2025</year><month>01</month><day>14</day><volume>27</volume><fpage>e60520</fpage><pub-id pub-id-type="doi">10.2196/60520</pub-id><pub-id pub-id-type="medline">39576242</pub-id></nlm-citation></ref><ref id="ref4"><label>4</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Eadie</surname><given-names>P</given-names> </name><name name-style="western"><surname>Morgan</surname><given-names>A</given-names> </name><name name-style="western"><surname>Ukoumunne</surname><given-names>OC</given-names> </name><name name-style="western"><surname>Ttofari Eecen</surname><given-names>K</given-names> </name><name name-style="western"><surname>Wake</surname><given-names>M</given-names> </name><name name-style="western"><surname>Reilly</surname><given-names>S</given-names> </name></person-group><article-title>Speech sound disorder at 4 years: prevalence, comorbidities, and predictors in a community cohort of children</article-title><source>Dev Med Child Neurol</source><year>2015</year><month>06</month><volume>57</volume><issue>6</issue><fpage>578</fpage><lpage>584</lpage><pub-id pub-id-type="doi">10.1111/dmcn.12635</pub-id><pub-id pub-id-type="medline">25403868</pub-id></nlm-citation></ref><ref id="ref5"><label>5</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Peterson</surname><given-names>RL</given-names> </name><name name-style="western"><surname>Pennington</surname><given-names>BF</given-names> </name><name name-style="western"><surname>Shriberg</surname><given-names>LD</given-names> </name><name name-style="western"><surname>Boada</surname><given-names>R</given-names> </name></person-group><article-title>What influences literacy outcome in children with speech sound disorder?</article-title><source>J Speech Lang Hear Res</source><year>2009</year><month>10</month><volume>52</volume><issue>5</issue><fpage>1175</fpage><lpage>1188</lpage><pub-id pub-id-type="doi">10.1044/1092-4388(2009/08-0024)</pub-id><pub-id pub-id-type="medline">19403946</pub-id></nlm-citation></ref><ref id="ref6"><label>6</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Singh</surname><given-names>L</given-names> </name><name name-style="western"><surname>Tan</surname><given-names>A</given-names> </name><name name-style="western"><surname>Quinn</surname><given-names>PC</given-names> </name></person-group><article-title>Infants recognize words spoken through opaque masks but not through clear masks</article-title><source>Dev Sci</source><year>2021</year><month>11</month><volume>24</volume><issue>6</issue><fpage>e13117</fpage><pub-id pub-id-type="doi">10.1111/desc.13117</pub-id><pub-id pub-id-type="medline">33942441</pub-id></nlm-citation></ref><ref id="ref7"><label>7</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Pejovic</surname><given-names>J</given-names> </name><name name-style="western"><surname>Severino</surname><given-names>C</given-names> </name><name name-style="western"><surname>Vig&#x00E1;rio</surname><given-names>M</given-names> </name><name name-style="western"><surname>Frota</surname><given-names>S</given-names> </name></person-group><article-title>Prolonged COVID-19 related effects on early language development: a longitudinal study</article-title><source>Early Hum Dev</source><year>2024</year><month>08</month><volume>195</volume><fpage>106081</fpage><pub-id pub-id-type="doi">10.1016/j.earlhumdev.2024.106081</pub-id><pub-id pub-id-type="medline">39047632</pub-id></nlm-citation></ref><ref id="ref8"><label>8</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Qvarfordt</surname><given-names>M</given-names> </name><name name-style="western"><surname>Nilsson</surname><given-names>E</given-names> </name><name name-style="western"><surname>Nilsson</surname><given-names>L</given-names> </name></person-group><article-title>Health care professionals&#x2019; experiences in telerehabilitation: qualitative content analysis</article-title><source>JMIR Hum Factors</source><year>2023</year><month>04</month><day>19</day><volume>10</volume><fpage>e40690</fpage><pub-id pub-id-type="doi">10.2196/40690</pub-id><pub-id pub-id-type="medline">37074772</pub-id></nlm-citation></ref><ref id="ref9"><label>9</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Nguyen</surname><given-names>G</given-names> </name><name name-style="western"><surname>King</surname><given-names>K</given-names> </name><name name-style="western"><surname>Stirling</surname><given-names>L</given-names> </name></person-group><article-title>Telerehabilitation use and experiences in occupational and physical therapy through the early stages of the COVID-19 pandemic</article-title><source>PLoS One</source><year>2023</year><volume>18</volume><issue>11</issue><fpage>e0291605</fpage><pub-id pub-id-type="doi">10.1371/journal.pone.0291605</pub-id><pub-id pub-id-type="medline">37939089</pub-id></nlm-citation></ref><ref id="ref10"><label>10</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Park</surname><given-names>ES</given-names> </name><name name-style="western"><surname>Lee</surname><given-names>EJ</given-names> </name><name name-style="western"><surname>Lee</surname><given-names>R</given-names> </name></person-group><article-title>Accuracy of final consonants at word-medial position according to percentage of correct consonants in children with speech sound disorders</article-title><source>J Speech Lang Hear Disord</source><year>2017</year><volume>26</volume><issue>1</issue><fpage>77</fpage><lpage>89</lpage><pub-id pub-id-type="doi">10.15724/jslhd.2017.26.1.008</pub-id></nlm-citation></ref><ref id="ref11"><label>11</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Baumel</surname><given-names>A</given-names> </name></person-group><article-title>More than a chatbot: a practical framework to harness artificial intelligence across key components to boost digital therapeutics quality</article-title><source>Front Digit Health</source><year>2025</year><volume>7</volume><fpage>1541676</fpage><pub-id pub-id-type="doi">10.3389/fdgth.2025.1541676</pub-id><pub-id pub-id-type="medline">40343211</pub-id></nlm-citation></ref><ref id="ref12"><label>12</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Dang</surname><given-names>A</given-names> </name><name name-style="western"><surname>Arora</surname><given-names>D</given-names> </name><name name-style="western"><surname>Rane</surname><given-names>P</given-names> </name></person-group><article-title>Role of digital therapeutics and the changing future of healthcare</article-title><source>J Family Med Prim Care</source><year>2020</year><month>05</month><volume>9</volume><issue>5</issue><fpage>2207</fpage><lpage>2213</lpage><pub-id pub-id-type="doi">10.4103/jfmpc.jfmpc_105_20</pub-id><pub-id pub-id-type="medline">32754475</pub-id></nlm-citation></ref><ref id="ref13"><label>13</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Phan</surname><given-names>P</given-names> </name><name name-style="western"><surname>Mitragotri</surname><given-names>S</given-names> </name><name name-style="western"><surname>Zhao</surname><given-names>Z</given-names> </name></person-group><article-title>Digital therapeutics in the clinic</article-title><source>Bioeng Transl Med</source><year>2023</year><month>05</month><volume>8</volume><issue>4</issue><fpage>e10536</fpage><pub-id pub-id-type="doi">10.1002/btm2.10536</pub-id><pub-id pub-id-type="medline">37476062</pub-id></nlm-citation></ref><ref id="ref14"><label>14</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Sung</surname><given-names>SS</given-names> </name><name name-style="western"><surname>So</surname><given-names>J</given-names> </name><name name-style="western"><surname>Yoon</surname><given-names>TJ</given-names> </name><name name-style="western"><surname>Ha</surname><given-names>S</given-names> </name></person-group><article-title>Automatic detection of speech sound disorder in children using automatic speech recognition and audio classification</article-title><source>Phon Speech Sci</source><year>2024</year><month>09</month><volume>16</volume><issue>3</issue><fpage>87</fpage><lpage>94</lpage><pub-id pub-id-type="doi">10.13064/KSSS.2024.16.3.087</pub-id></nlm-citation></ref><ref id="ref15"><label>15</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Hirsch</surname><given-names>ME</given-names> </name><name name-style="western"><surname>Thompson</surname><given-names>A</given-names> </name><name name-style="western"><surname>Kim</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Lansford</surname><given-names>KL</given-names> </name></person-group><article-title>The reliability and validity of speech-language pathologists&#x2019; estimations of intelligibility in dysarthria</article-title><source>Brain Sci</source><year>2022</year><month>07</month><day>30</day><volume>12</volume><issue>8</issue><fpage>1011</fpage><pub-id pub-id-type="doi">10.3390/brainsci12081011</pub-id><pub-id pub-id-type="medline">36009074</pub-id></nlm-citation></ref><ref id="ref16"><label>16</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Boulkedid</surname><given-names>R</given-names> </name><name name-style="western"><surname>Abdoul</surname><given-names>H</given-names> </name><name name-style="western"><surname>Loustau</surname><given-names>M</given-names> </name><name name-style="western"><surname>Sibony</surname><given-names>O</given-names> </name><name name-style="western"><surname>Alberti</surname><given-names>C</given-names> </name></person-group><article-title>Using and reporting the Delphi method for selecting healthcare quality indicators: a systematic review</article-title><source>PLoS One</source><year>2011</year><volume>6</volume><issue>6</issue><fpage>e20476</fpage><pub-id pub-id-type="doi">10.1371/journal.pone.0020476</pub-id><pub-id pub-id-type="medline">21694759</pub-id></nlm-citation></ref><ref id="ref17"><label>17</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Varghese</surname><given-names>AL</given-names> </name><name name-style="western"><surname>Balasubramanium</surname><given-names>RK</given-names> </name><name name-style="western"><surname>Bajaj</surname><given-names>G</given-names> </name><name name-style="western"><surname>Karuppali</surname><given-names>S</given-names> </name><name name-style="western"><surname>Bhaskaran</surname><given-names>U</given-names> </name></person-group><article-title>Delphi survey of items for the Test of Stuttering Screening in Children (TSSC)</article-title><source>Clin Pract Epidemiol Ment Health</source><year>2023</year><volume>19</volume><fpage>e174501792305170</fpage><pub-id pub-id-type="doi">10.2174/17450179-v19-230615-2022-27</pub-id><pub-id pub-id-type="medline">37916211</pub-id></nlm-citation></ref><ref id="ref18"><label>18</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Clayton</surname><given-names>MJ</given-names> </name></person-group><article-title>Delphi: a technique to harness expert opinion for critical decision&#x2010;making tasks in education</article-title><source>Educ Psychol</source><year>1997</year><volume>17</volume><issue>4</issue><fpage>373</fpage><lpage>386</lpage><pub-id pub-id-type="doi">10.1080/0144341970170401</pub-id></nlm-citation></ref><ref id="ref19"><label>19</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Ayre</surname><given-names>C</given-names> </name><name name-style="western"><surname>Scally</surname><given-names>AJ</given-names> </name></person-group><article-title>Critical values for Lawshe&#x2019;s content validity ratio: revisiting the original methods of calculation</article-title><source>Meas Eval Couns Dev</source><year>2014</year><volume>47</volume><issue>1</issue><fpage>79</fpage><lpage>86</lpage><pub-id pub-id-type="doi">10.1177/0748175613513808</pub-id></nlm-citation></ref><ref id="ref20"><label>20</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Brooks</surname><given-names>D</given-names> </name><name name-style="western"><surname>Sriram</surname><given-names>D</given-names> </name><name name-style="western"><surname>Brimelow</surname><given-names>R</given-names> </name><etal/></person-group><article-title>Mental health-related quality indicators for long-term care: a modified Delphi consensus study</article-title><source>J Am Med Dir Assoc</source><year>2025</year><month>09</month><volume>26</volume><issue>9</issue><fpage>105768</fpage><pub-id pub-id-type="doi">10.1016/j.jamda.2025.105768</pub-id><pub-id pub-id-type="medline">40652957</pub-id></nlm-citation></ref><ref id="ref21"><label>21</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Mansilla</surname><given-names>C</given-names> </name><name name-style="western"><surname>Guyatt</surname><given-names>G</given-names> </name><name name-style="western"><surname>Sweetman</surname><given-names>A</given-names> </name><name name-style="western"><surname>Lavis</surname><given-names>JN</given-names> </name></person-group><article-title>Matching the right study design to decision-maker questions: results from a Delphi study</article-title><source>PLOS Glob Public Health</source><year>2024</year><volume>4</volume><issue>2</issue><fpage>e0002752</fpage><pub-id pub-id-type="doi">10.1371/journal.pgph.0002752</pub-id><pub-id pub-id-type="medline">38421991</pub-id></nlm-citation></ref><ref id="ref22"><label>22</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>MacNeela</surname><given-names>P</given-names> </name><name name-style="western"><surname>Morris</surname><given-names>R</given-names> </name><name name-style="western"><surname>Scott</surname><given-names>A</given-names> </name><name name-style="western"><surname>Treacy</surname><given-names>MP</given-names> </name><name name-style="western"><surname>Hyde</surname><given-names>A</given-names> </name></person-group><article-title>Seen as core: a Delphi consensus study of essential elements of mental health nursing care in Ireland</article-title><source>Arch Psychiatr Nurs</source><year>2010</year><month>10</month><volume>24</volume><issue>5</issue><fpage>339</fpage><lpage>348</lpage><pub-id pub-id-type="doi">10.1016/j.apnu.2010.02.001</pub-id><pub-id pub-id-type="medline">20851325</pub-id></nlm-citation></ref></ref-list><app-group><supplementary-material id="app1"><label>Multimedia Appendix 1</label><p>Round 1-3 Delphi questionnaire.</p><media xlink:href="formative_v10i1e90110_app1.docx" xlink:title="DOCX File, 48 KB"/></supplementary-material><supplementary-material id="app2"><label>Checklist 1</label><p>DELPHISTAR checklist.</p><media xlink:href="formative_v10i1e90110_app2.pdf" xlink:title="PDF File, 798 KB"/></supplementary-material></app-group></back></article>