<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.0 20040830//EN" "http://dtd.nlm.nih.gov/publishing/2.0/journalpublishing.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink" article-type="research-article" dtd-version="2.0">
  <front>
    <journal-meta>
      <journal-id journal-id-type="publisher-id">JFR</journal-id>
      <journal-id journal-id-type="nlm-ta">JMIR Form Res</journal-id>
      <journal-title>JMIR Formative Research</journal-title>
      <issn pub-type="epub">2561-326X</issn>
      <publisher>
        <publisher-name>JMIR Publications</publisher-name>
        <publisher-loc>Toronto, Canada</publisher-loc>
      </publisher>
    </journal-meta>
    <article-meta>
      <article-id pub-id-type="publisher-id">v6i10e39998</article-id>
      <article-id pub-id-type="pmid">36306165</article-id>
      <article-id pub-id-type="doi">10.2196/39998</article-id>
      <article-categories>
        <subj-group subj-group-type="heading">
          <subject>Original Paper</subject>
        </subj-group>
        <subj-group subj-group-type="article-type">
          <subject>Original Paper</subject>
        </subj-group>
      </article-categories>
      <title-group>
        <article-title>Screening for Generalized Anxiety Disorder From Acoustic and Linguistic Features of Impromptu Speech: Prediction Model Evaluation Study</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="editor">
          <name>
            <surname>Mavragani</surname>
            <given-names>Amaryllis</given-names>
          </name>
        </contrib>
      </contrib-group>
      <contrib-group>
        <contrib contrib-type="reviewer">
          <name>
            <surname>Teachman</surname>
            <given-names>Bethany</given-names>
          </name>
        </contrib>
        <contrib contrib-type="reviewer">
          <name>
            <surname>Corcoran</surname>
            <given-names>Cheryl</given-names>
          </name>
        </contrib>
      </contrib-group>
      <contrib-group>
        <contrib id="contrib1" contrib-type="author" corresp="yes">
          <name name-style="western">
            <surname>Teferra</surname>
            <given-names>Bazen Gashaw</given-names>
          </name>
          <degrees>BSc, MSc</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <address>
            <institution>The Edward S Rogers Sr Department of Electrical and Computer Engineering</institution>
            <institution>University of Toronto</institution>
            <addr-line>10 King’s College Road</addr-line>
            <addr-line>Toronto, ON, M5S 3G4</addr-line>
            <country>Canada</country>
            <phone>1 4169786992</phone>
            <email>bazen.teferra@mail.utoronto.ca</email>
          </address>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0001-5325-9639</ext-link>
        </contrib>
        <contrib id="contrib2" contrib-type="author">
          <name name-style="western">
            <surname>Borwein</surname>
            <given-names>Sophie</given-names>
          </name>
          <degrees>BA, MPP, PhD</degrees>
          <xref rid="aff2" ref-type="aff">2</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0002-6698-6648</ext-link>
        </contrib>
        <contrib id="contrib3" contrib-type="author">
          <name name-style="western">
            <surname>DeSouza</surname>
            <given-names>Danielle D</given-names>
          </name>
          <degrees>MSc, PhD</degrees>
          <xref rid="aff3" ref-type="aff">3</xref>
          <xref rid="aff4" ref-type="aff">4</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0001-6861-5691</ext-link>
        </contrib>
        <contrib id="contrib4" contrib-type="author">
          <name name-style="western">
            <surname>Rose</surname>
            <given-names>Jonathan</given-names>
          </name>
          <degrees>BASc, MASc, PhD</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <xref rid="aff5" ref-type="aff">5</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0002-3551-2175</ext-link>
        </contrib>
      </contrib-group>
      <aff id="aff1">
        <label>1</label>
        <institution>The Edward S Rogers Sr Department of Electrical and Computer Engineering</institution>
        <institution>University of Toronto</institution>
        <addr-line>Toronto, ON</addr-line>
        <country>Canada</country>
      </aff>
      <aff id="aff2">
        <label>2</label>
        <institution>School of Public Policy</institution>
        <institution>Simon Fraser University</institution>
        <addr-line>Vancouver, BC</addr-line>
        <country>Canada</country>
      </aff>
      <aff id="aff3">
        <label>3</label>
        <institution>Winterlight Labs</institution>
        <addr-line>Toronto, ON</addr-line>
        <country>Canada</country>
      </aff>
      <aff id="aff4">
        <label>4</label>
        <institution>Department of Neurology and Neurological Sciences</institution>
        <institution>Stanford University</institution>
        <addr-line>Palo Alto, CA</addr-line>
        <country>United States</country>
      </aff>
      <aff id="aff5">
        <label>5</label>
        <institution>The Centre for Addiction and Mental Health</institution>
        <addr-line>Toronto, ON</addr-line>
        <country>Canada</country>
      </aff>
      <author-notes>
        <corresp>Corresponding Author: Bazen Gashaw Teferra <email>bazen.teferra@mail.utoronto.ca</email></corresp>
      </author-notes>
      <pub-date pub-type="collection">
        <month>10</month>
        <year>2022</year>
      </pub-date>
      <pub-date pub-type="epub">
        <day>28</day>
        <month>10</month>
        <year>2022</year>
      </pub-date>
      <volume>6</volume>
      <issue>10</issue>
      <elocation-id>e39998</elocation-id>
      <history>
        <date date-type="received">
          <day>1</day>
          <month>6</month>
          <year>2022</year>
        </date>
        <date date-type="rev-request">
          <day>10</day>
          <month>8</month>
          <year>2022</year>
        </date>
        <date date-type="rev-recd">
          <day>29</day>
          <month>9</month>
          <year>2022</year>
        </date>
        <date date-type="accepted">
          <day>30</day>
          <month>9</month>
          <year>2022</year>
        </date>
      </history>
      <copyright-statement>©Bazen Gashaw Teferra, Sophie Borwein, Danielle D DeSouza, Jonathan Rose. Originally published in JMIR Formative Research (https://formative.jmir.org), 28.10.2022.</copyright-statement>
      <copyright-year>2022</copyright-year>
      <license license-type="open-access" xlink:href="https://creativecommons.org/licenses/by/4.0/">
        <p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR Formative Research, is properly cited. The complete bibliographic information, a link to the original publication on https://formative.jmir.org, as well as this copyright and license information must be included.</p>
      </license>
      <self-uri xlink:href="https://formative.jmir.org/2022/10/e39998" xlink:type="simple"/>
      <abstract>
        <sec sec-type="background">
          <title>Background</title>
          <p>Frequent interaction with mental health professionals is required to screen, diagnose, and track mental health disorders. However, high costs and insufficient access can make frequent interactions difficult. The ability to assess a mental health disorder passively and at frequent intervals could be a useful complement to the conventional treatment. It may be possible to passively assess clinical symptoms with high frequency by characterizing speech alterations collected using personal smartphones or other wearable devices. The association between speech features and mental health disorders can be leveraged as an objective screening tool.</p>
        </sec>
        <sec sec-type="objective">
          <title>Objective</title>
          <p>This study aimed to evaluate the performance of a model that predicts the presence of generalized anxiety disorder (GAD) from acoustic and linguistic features of impromptu speech on a larger and more generalizable scale than prior studies did.</p>
        </sec>
        <sec sec-type="methods">
          <title>Methods</title>
          <p>A total of 2000 participants were recruited, and they participated in a single web-based session. They completed the Generalized Anxiety Disorder-7 item scale assessment and provided an impromptu speech sample in response to a modified version of the Trier Social Stress Test. We used the linguistic and acoustic features that were found to be associated with anxiety disorders in previous studies along with demographic information to predict whether participants fell above or below the screening threshold for GAD based on the Generalized Anxiety Disorder-7 item scale threshold of 10. Separate models for each sex were also evaluated. We reported the mean area under the receiver operating characteristic (AUROC) from a repeated 5-fold cross-validation to evaluate the performance of the models.</p>
        </sec>
        <sec sec-type="results">
          <title>Results</title>
          <p>A logistic regression model using only acoustic and linguistic speech features achieved a significantly greater prediction accuracy than a random model did (mean AUROC 0.57, SD 0.03; <italic>P&#60;.</italic>001). When separately assessing samples from female participants, we observed a mean AUROC of 0.55 (SD 0.05; <italic>P</italic>=.01). The model constructed from the samples from male participants achieved a mean AUROC of 0.57 (SD 0.07; <italic>P=.</italic>002). The mean AUROC increased to 0.62 (SD 0.03; <italic>P&#60;.</italic>001) on the all-sample data set when demographic information (age, sex, and income) was included, indicating the importance of demographics when screening for anxiety disorders. The performance also increased for the female sample to a mean of 0.62 (SD 0.04; <italic>P&#60;.</italic>001) when using demographic information (age and income). An increase in performance was not observed when demographic information was added to the model constructed from the male samples.</p>
        </sec>
        <sec sec-type="conclusions">
          <title>Conclusions</title>
          <p>A logistic regression model using acoustic and linguistic speech features, which have been suggested to be associated with anxiety disorders in prior studies, can achieve above-random accuracy for predicting GAD. Importantly, the addition of basic demographic variables further improves model performance, suggesting a role for speech and demographic information to be used as automated, objective screeners of GAD.</p>
        </sec>
      </abstract>
      <kwd-group>
        <kwd>mental health</kwd>
        <kwd>generalized anxiety disorder</kwd>
        <kwd>impromptu speech</kwd>
        <kwd>acoustic features</kwd>
        <kwd>linguistic features</kwd>
        <kwd>anxiety prediction</kwd>
        <kwd>mobile phone</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec sec-type="introduction">
      <title>Introduction</title>
      <sec>
        <title>Background</title>
        <p>Anxiety disorders are characterized by an excessive and uncontrollable fear of what is to come and are associated with preparation for possible future adverse events [<xref ref-type="bibr" rid="ref1">1</xref>]. Although anxiety is an important emotion that helps us prepare for future events, it limits the performance of day-to-day tasks when it becomes uncontrollable. Anxiety disorders are one of the most common mental health issues with an incidence of approximately 10% in the Canadian population [<xref ref-type="bibr" rid="ref2">2</xref>]. Unfortunately, many Canadians affected by anxiety are unable to access psychological and psychiatric resources [<xref ref-type="bibr" rid="ref3">3</xref>] due in part to the cost [<xref ref-type="bibr" rid="ref4">4</xref>] and the general lack of availability [<xref ref-type="bibr" rid="ref5">5</xref>]. Some of this deficit may be addressed using methods that automate certain aspects of the measurement and diagnosis of anxiety disorders.</p>
        <p>In this study, we focused on generalized anxiety disorder (GAD) [<xref ref-type="bibr" rid="ref6">6</xref>] and sought to automatically detect GAD from speech. We monitored speech because it is possible to passively and frequently sample ambient speech, ensuring that the privacy and confidentiality of the participants are appropriately handled. The capability to detect anxiety from ambient speech could be part of a system to automatically screen for anxiety, monitor treatment, and detect relapse.</p>
        <p>We anticipated the following scenario for a system that included the capability to automatically predict anxiety from speech. Such a system would sample a sequence of the participant’s speech throughout the day and produce multiple predictions. Depending on the accuracy of individual predictions, the system could use multiple predictions to increase the overall accuracy of the final screening result. Note that this approach uses passively collected speech, which gives rise to its own challenges, including the need for a process of speaker identification to select words spoken by the participant.</p>
        <p>Another motivation for pursuing the automatic detection of anxiety from speech is to avoid the subjectivity normally present in the screening and diagnosis of GAD. The current gold standard diagnosis for GAD is influenced by both the subject information supplied by a patient to a clinician and the subjective judgment of that information by the clinician. This subjectivity can lead to inaccurate diagnostic outcomes in patients [<xref ref-type="bibr" rid="ref7">7</xref>]. There is a potential benefit of an objective marker of anxiety. In this study, we explored how well such a biomarker could be obtained from a person’s speech. Prior studies suggest that anxiety influences the acoustic features of speech [<xref ref-type="bibr" rid="ref8">8</xref>], as these features are difficult for a person to control [<xref ref-type="bibr" rid="ref9">9</xref>]. Moreover, anxiety may also manifest in the choice of words, which we refer to as linguistic features [<xref ref-type="bibr" rid="ref10">10</xref>].</p>
        <p>In an earlier study [<xref ref-type="bibr" rid="ref11">11</xref>], we identified acoustic and linguistic features of speech that were correlated with anxiety as measured by the Generalized Anxiety Disorder 7-item (GAD-7) scale. Building on our earlier study and using the same participants and data, in this study, we aimed to build and measure the performance of a model that predicts whether participants are above or below the screening threshold for GAD on a much larger scale than what prior studies did. Previous studies validating the GAD-7 scale have shown that using a cut point of ≥10 optimizes sensitivity and specificity for identifying individuals with a diagnosis of GAD [<xref ref-type="bibr" rid="ref12">12</xref>]. The model makes use of the previously identified features of speech together with the demographics of the participants.</p>
        <p>This paper is organized as follows: the next subsection summarizes related work on anxiety prediction and proposes a hypothesis. The Methods section describes the speech sample collection methods, set of features used, and construction and evaluation of predictive models. The Results section presents the demographics of the participants and performance of the prediction model, while the Discussion section discusses the results and their implications for future research on anxiety detection.</p>
      </sec>
      <sec>
        <title>Related Work</title>
        <p>Several previous studies have measured the association between speech features and various forms of anxiety, and other efforts sought to automatically detect anxiety from acoustic and linguistic features of speech. The studies explored in this section examine a broader class of anxiety disorders, including internalizing disorders, social phobia or social anxiety disorders (SADs), panic disorder, agoraphobia, and GAD.</p>
        <p>McGinnis et al [<xref ref-type="bibr" rid="ref13">13</xref>] identified several acoustic characteristics of speech that can be used to detect anxiety disorders in children. They studied the speech of 71 participants between the ages of 3 and 8 years and were able to detect internalizing disorders (a collective term for anxiety and depression). The authors extracted and selected several acoustic features from the speech produced in a 3-minute task based on the Trier Social Stress Test (TSST) for children [<xref ref-type="bibr" rid="ref14">14</xref>]. These features included the zero-crossing rate, Mel frequency cepstral coefficients [<xref ref-type="bibr" rid="ref15">15</xref>], zero-crossing rate of the <italic>z</italic> score of the power spectral density, dominant frequency, mean frequency, perceptual spectral centroid, spectral flatness, and the skew and kurtosis of the power spectral density. Several models were built to predict whether children had an internalizing disorder (43/71, 61%) or were healthy. Both logistic regression (LR) and a support vector machine (SVM) analyses [<xref ref-type="bibr" rid="ref16">16</xref>] achieved a classification accuracy of 80%.</p>
        <p>Weeks et al [<xref ref-type="bibr" rid="ref17">17</xref>] found a relationship between anxiety and voice alterations. Their study showed a link between vocal pitch (characterized by fundamental frequency [F0]) and SAD. They collected impromptu speech samples from 46 undergraduate students, 25 (54%) with a diagnosis of SAD and 21 (46%) healthy controls. The participants also completed the Beck Anxiety Scale as a measure of self-reported anxiety severity [<xref ref-type="bibr" rid="ref18">18</xref>]. Their results indicated that the mean F0 was positively correlated (<italic>r</italic>=0.72; <italic>P</italic>=.002) with anxiety severity in all male participants. However, the correlation in female participants was weaker (<italic>r</italic>=0.02; <italic>P</italic>=.92), indicating possible sex differences in the relationship between anxiety severity and vocal pitch. In a related continuing study [<xref ref-type="bibr" rid="ref19">19</xref>], the authors attempted to classify men with SAD using the mean F0. Using a mean F0 value of 122.78 Hz, they achieved a sensitivity of 89% (8/9 male patients with SAD correctly classified) and a specificity of 100% (4/4 male healthy controls correctly classified).</p>
        <p>Salekin et al [<xref ref-type="bibr" rid="ref20">20</xref>] explored methods for detecting social anxiety and depression from an audio clip of a person’s speech. Their data set included a 3-minutes speech sample from each of the 105 participants describing what they liked and disliked about college or their hometown. The participants were asked to report their peak levels of anxiety during the speech. The authors presented and used a novel feature modeling technique called NN2Vec that can identify the relationship between a participant’s speech and affective states. Using the features from NN2Vec and a bidirectional Long Short-Term Memory Multiple Instance Learning network, they were able to detect speakers with high social anxiety with an <italic>F</italic><sub>1</sub>-score of 90.1% and speakers with depression symptoms with an <italic>F</italic><sub>1</sub>-score of 85.4%.</p>
        <p>Baird et al [<xref ref-type="bibr" rid="ref21">21</xref>] explored the effect of anxiety on speech by attempting to predict anxiety using sustained vowels. Their data set comprised 239 speakers (69 male participants) aged 18 to 68 years who performed various vocal exercises, which included sustained vowel sounds. They used the Beck Anxiety Inventory (BAI) [<xref ref-type="bibr" rid="ref22">22</xref>] questionnaire as a label for each participant. The BAI is also one of the scales used to screen for GAD. They used 4 classes of sustained <italic>(a)</italic> vowels from each participant: a sad phonation, a smiling phonation, a comfortable phonation, and a powerful or loud phonation. From the sustained vowels, they extracted acoustic features such as the SD of F0, intensity, and harmonic-to-noise ratio. Using these features and a BAI label, they trained a support vector regressor with a linear kernel and used Spearman correlation between the predicted and the actual label to evaluate the performance of their model. They split their data into training and test sets and achieved a Spearman correlation of 0.243 on the test data set. They reported a better performance of a Spearman correlation of 0.59 when they only considered the group with high BAI scores, indicating that the symptoms of anxiety are more observable in individuals with high anxiety.</p>
        <p>Rook et al [<xref ref-type="bibr" rid="ref23">23</xref>] hypothesized that the worrying behavior in GAD comes from the verbal linguistic process. They attempted to predict GAD using only linguistic patterns. A total of 142 undergraduate students (56 male and 86 female participants) were recruited and asked to recall and write down an anxious experience during their university life. Each participant filled out the GAD-7 scale score and the behavioral inhibition/behavioral approach scale (BIS/BAS) [<xref ref-type="bibr" rid="ref24">24</xref>]. The Linguistic Inquiry and Word Count (LIWC) [<xref ref-type="bibr" rid="ref25">25</xref>] method was used to extract features from the texts written by the participants. Another set of features was also used by combining the LIWC features with BIS/BAS scores. Several machine learning models were explored, including SVM with linear kernel, LR, naïve Bayes, and random forest. Their results showed that all the models performed significantly better than a random model. In addition, better performance was obtained from all the models except the SVM when the LIWC and BIS/BAS features were used together as inputs compared with using only the LIWC features.</p>
        <p>Di Matteo et al [<xref ref-type="bibr" rid="ref26">26</xref>] examined the relationship between passively collected audio data and anxiety and depression. Their study continued for 2 weeks, where 84 participants installed an app on their smartphone that collected the average volume of sounds (the average of 15-second audio collected every 5 minutes) and the presence or absence of speech in the environment. They then extracted 4 environmental audio-based features: daily similarity, sleep disturbance (on all nights and weeknights only), and speech-presence ratio. Their results showed that none of the extracted features were significantly correlated with anxiety. However, these features were significantly correlated with depression: daily similarities (<italic>r</italic>=−0.37; <italic>P</italic>&#60;.001), sleep disturbance on weeknights (<italic>r</italic>=0.23; <italic>P</italic>=.03), and speech presence (<italic>r</italic>=−0.37; <italic>P</italic>&#60;.001).</p>
        <p>Di Matteo et al [<xref ref-type="bibr" rid="ref27">27</xref>] also explored the relationship between linguistic features of speech and anxiety. They used passively collected intermittent samples of audio data from participants’ smartphones, which they converted to text. The authors used the LIWC approach [<xref ref-type="bibr" rid="ref25">25</xref>] to classify words into 67 categories. They calculated correlations using 4 self-report measures: SAD, GAD, depression, and functional impairment. They observed a significant correlation between words related to perceptual process (<italic>See</italic> in the LIWC) with SAD (<italic>r</italic>=0.31; <italic>P</italic>=.003) and words related to rewards with GAD (<italic>r</italic>=−0.29; <italic>P</italic>=.007).</p>
        <p>In their third study, using the data collected from the 84 participants, Di Matteo et al [<xref ref-type="bibr" rid="ref28">28</xref>] attempted to predict GAD, SAD, and depression from the smartphone-collected data. The features used in this study included daily similarity, speech presence, weeknight sleep disturbance, death-related words, number of locations visited, number of exits from home, screen use, and time in darkness. Although the models built on these features achieved an above-random prediction accuracy for SAD and depression, they did not observe above-random prediction accuracy for GAD.</p>
        <p>Overall, prior studies suggest that it is possible to detect anxiety disorders from speech. However, the largest sample size among these previous studies was a total of 239, with an average of 115 participants, which limits the generalizability of the results. In addition, the number of participants might not be the only factor affecting generalizability. Apart from the studies by Di Matteo et al [<xref ref-type="bibr" rid="ref28">28</xref>] and Baird et al [<xref ref-type="bibr" rid="ref21">21</xref>], the prior studies were mostly limited to very specific demographics: McGinnis et al [<xref ref-type="bibr" rid="ref13">13</xref>] focused on children; Weeks et al [<xref ref-type="bibr" rid="ref17">17</xref>], Salekin et al [<xref ref-type="bibr" rid="ref20">20</xref>], and Rook et al [<xref ref-type="bibr" rid="ref23">23</xref>] focused on undergraduate students at a university or college.</p>
        <p>We hypothesized that by recruiting a substantially larger cohort (N=2000) with broader demographic characteristics than that in prior studies, it is possible to achieve above-random prediction accuracy in screening for GAD using acoustic and linguistic features that have been previously suggested.</p>
      </sec>
    </sec>
    <sec sec-type="methods">
      <title>Methods</title>
      <sec>
        <title>Data Collection</title>
        <sec>
          <title>Recruitment and Demographics</title>
          <p>We must note that the participants recruited and the data used in this study are the same as those in our earlier study [<xref ref-type="bibr" rid="ref11">11</xref>], which focused solely on the correlations between acoustic and linguistic features of speech and the GAD-7. This study used those features and additional demographics to construct a predictive model. Participants were recruited from a nonclinical population using Prolific [<xref ref-type="bibr" rid="ref29">29</xref>], a web-based human participant recruitment platform. The inclusion criteria for this study were an age range of 18 to 65 years, fluency in English, English as a first language, and at least 10 previous studies completed on Prolific, with 95% of these previous Prolific tasks completed satisfactorily (as labeled by the study author). The Prolific platform also provided several relevant demographics of the participants, including their age and income. The data set was also balanced for sex (50% female and 50% male).</p>
          <p>Participants who completed the study were paid £2 (approximately CAD $3.41; US $2.74) for approximately 15 minutes of work. They completed the entire study remotely, using their PCs.</p>
        </sec>
        <sec>
          <title>Study Procedure</title>
          <p>Participants were recruited on Prolific for a 10- to 15-minute task implemented through a custom website. Our earlier paper on the correlates of anxiety [<xref ref-type="bibr" rid="ref11">11</xref>] described the data collection procedure in detail. Parts of the data collection procedure that are relevant for the purpose of this study are described in the following sections.</p>
          <p>On the Prolific platform, individuals who met the inclusion criteria were presented with the opportunity to participate in this study. Those who wished to participate clicked on the study link, which brought them to a consent form that described the procedure and goals of the study and provided information on data privacy. If a participant provided consent, a hyperlink brought them to an external web app that implemented the tasks described in further sections.</p>
          <p>Participants were asked to fill out the standard GAD-7 questionnaire [<xref ref-type="bibr" rid="ref12">12</xref>] described in more detail in the Anxiety Measures section. They were then asked to perform a speech task, which was recorded using their computer microphone. The speech task followed a modified version of the widely used TSST [<xref ref-type="bibr" rid="ref30">30</xref>], which aimed to evoke a moderate amount of stress from each participant. Prior studies [<xref ref-type="bibr" rid="ref31">31</xref>,<xref ref-type="bibr" rid="ref32">32</xref>] have shown higher activation (cardiovascular, skin conductance, and plasma levels of norepinephrine and testosterone) in participants with relatively higher anxiety after exposure to moderate stress induced by the TSST.</p>
          <p>In the modified version of TSST, participants were told to imagine that they were job applicants invited for an interview with a hiring manager. They were told to imagine that it was a job that they really wanted—their so-called “dream” job. They were given a few minutes to prepare—to choose their “dream” job—and to think about how they would convince an interviewer that they were the right person for that position. Participants were also told that the recorded video would be viewed by researchers studying their behavior and language. Participants were then asked to speak for 5 minutes, making the case for themselves to be hired for that dream job.</p>
          <p>Note that in the original TSST [<xref ref-type="bibr" rid="ref30">30</xref>], participants would normally deliver their speech in front of a live panel of judges. If a participant finished their delivery in &#60;5 minutes, the judges in the original TSST design would encourage the participant to speak for the full 5 minutes. For example, a statement of encouragement in the original TSST was, “What are your personal strengths?” In the modified TSST, we implemented a similar method to encourage participants to speak for the full 5 minutes; when our system detected silence (the absence of speech for &#62;6 seconds), it would display several different prompts inviting participants to keep speaking on different topics related to the task. Finally, the modified TSST only included the first part of the original TSST, not the second task, which involved the performance of mental arithmetic.</p>
        </sec>
        <sec>
          <title>Anxiety Measures</title>
          <p>We aimed to predict, based on features of the speech, if a participant is above or below the screening threshold for GAD based on the GAD-7 scale. The GAD-7 [<xref ref-type="bibr" rid="ref12">12</xref>] scale is a 7-item questionnaire that asks participants how often they were bothered by anxiety-related problems during the previous 2 weeks. Although the 2-week time period suggests that the GAD-7 measures a temporary condition, this is in contrast to the fact that a GAD diagnosis requires a 6-month duration of symptoms [<xref ref-type="bibr" rid="ref33">33</xref>,<xref ref-type="bibr" rid="ref34">34</xref>]. However, the GAD-7 has been validated as a diagnostic tool for GAD using a cutoff threshold of 10, with a sensitivity of 89% and specificity of 82% [<xref ref-type="bibr" rid="ref12">12</xref>]. Thus, we chose to use the GAD-7 threshold of 10 to obtain a binary label of GAD as the indicator of anxiety.</p>
          <p>Each of the 7 questions on the GAD-7 has 4 options for the participant to select, indicating how often they have been bothered by the 7 problems in the scale. These options and their numerical ratings are as follows: 0=not at all, 1=several days, 2=more than half the days, and 3=nearly every day. The final GAD-7 score is a summation of the values for each question, giving a severity measure for GAD ranging from 0 (no anxiety symptoms) to 21 (severe anxiety symptoms).</p>
        </sec>
      </sec>
      <sec>
        <title>Separation of Data for Analysis</title>
        <p>Certain demographic attributes were directly indicative of anxiety. For example, sex is known to influence the prevalence of anxiety [<xref ref-type="bibr" rid="ref35">35</xref>]. In addition, both age [<xref ref-type="bibr" rid="ref36">36</xref>] and income [<xref ref-type="bibr" rid="ref37">37</xref>] influence anxiety. Owing to the strong effect of sex and our interest in analyzing the effect of anxiety on both the sexes separately, we created a separate data set for female and male participants, in addition to the combined data set.</p>
      </sec>
      <sec>
        <title>Inputs to the Classification Model</title>
        <p>The inputs to our models were acoustic and linguistic features that were determined in a previous study [<xref ref-type="bibr" rid="ref11">11</xref>] to have a statistically significant correlation with the GAD-7. These features were found to be correlated with the GAD-7 after controlling for demographic variables such as age, sex, and personal income. These features are presented in <xref ref-type="table" rid="table1">Table 1</xref> as all-sample, female sample, and male sample data sets. The definitions of these features are presented in <xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref>.</p>
        <p>In addition to the acoustic and linguistic features, we explored the use of demographic information, such as age, sex, and personal income, as input features to the model. We decided to use these demographics as features in the model because they were available to us from the Prolific recruitment platform [<xref ref-type="bibr" rid="ref29">29</xref>]. Should the model be used as a diagnostic screener in the future, it should be possible to obtain these demographics.</p>
        <table-wrap position="float" id="table1">
          <label>Table 1</label>
          <caption>
            <p>Correlation of statistically significant acoustic and linguistic features with Generalized Anxiety Disorder-7 item (GAD-7) scale—results taken from earlier study (N=1744).</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="280"/>
            <col width="120"/>
            <col width="120"/>
            <col width="120"/>
            <col width="120"/>
            <col width="0"/>
            <col width="120"/>
            <col width="120"/>
            <thead>
              <tr valign="top">
                <td>Feature</td>
                <td colspan="7">Data set</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td colspan="2">All-sample</td>
                <td colspan="3">Female sample (n=862)</td>
                <td colspan="2">Male sample (n=882)</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>
                  <italic>r</italic>
                </td>
                <td><italic>P</italic> value</td>
                <td>
                  <italic>r</italic>
                </td>
                <td><italic>P</italic> value</td>
                <td colspan="2">
                  <italic>r</italic>
                </td>
                <td><italic>P</italic> value</td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td>AllPunc</td>
                <td>0.13</td>
                <td>&#60;.001</td>
                <td>0.14</td>
                <td>&#60;.001</td>
                <td colspan="2">0.13</td>
                <td>&#60;.001</td>
              </tr>
              <tr valign="top">
                <td>WC</td>
                <td>–0.12</td>
                <td>&#60;.001</td>
                <td>–0.13</td>
                <td>&#60;.001</td>
                <td colspan="2">–0.12</td>
                <td>&#60;.001</td>
              </tr>
              <tr valign="top">
                <td>Speaking duration</td>
                <td>–0.12</td>
                <td>&#60;.001</td>
                <td>–0.11</td>
                <td>&#60;.001</td>
                <td colspan="2">–0.13</td>
                <td>&#60;.001</td>
              </tr>
              <tr valign="top">
                <td>Period</td>
                <td>0.12</td>
                <td>&#60;.001</td>
                <td>0.16</td>
                <td>&#60;.001</td>
                <td colspan="2">0.08</td>
                <td>.02</td>
              </tr>
              <tr valign="top">
                <td>assent</td>
                <td>0.10</td>
                <td>&#60;.001</td>
                <td>0.10</td>
                <td>.004</td>
                <td colspan="2">0.11</td>
                <td>.001</td>
              </tr>
              <tr valign="top">
                <td>negemo</td>
                <td>0.10</td>
                <td>&#60;.001</td>
                <td>0.11</td>
                <td>&#60;.001</td>
                <td colspan="2">0.08</td>
                <td>.01</td>
              </tr>
              <tr valign="top">
                <td>relativ</td>
                <td>–0.09</td>
                <td>&#60;.001</td>
                <td>–0.09</td>
                <td>.006</td>
                <td colspan="2">–0.10</td>
                <td>.002</td>
              </tr>
              <tr valign="top">
                <td>motion</td>
                <td>–0.08</td>
                <td>&#60;.001</td>
                <td>–0.10</td>
                <td>.003</td>
                <td colspan="2">–0.07</td>
                <td>.048</td>
              </tr>
              <tr valign="top">
                <td>Shimmer</td>
                <td>0.08</td>
                <td>&#60;.001</td>
                <td>0.10</td>
                <td>.004</td>
                <td colspan="2">0.07</td>
                <td>.04</td>
              </tr>
              <tr valign="top">
                <td>swear</td>
                <td>0.08</td>
                <td>&#60;.001</td>
                <td>—<sup>a</sup></td>
                <td>—</td>
                <td colspan="2">0.10</td>
                <td>.004</td>
              </tr>
              <tr valign="top">
                <td>anger</td>
                <td>0.08</td>
                <td>&#60;.001</td>
                <td>0.11</td>
                <td>.002</td>
                <td colspan="2">—</td>
                <td>—</td>
              </tr>
              <tr valign="top">
                <td>mfcc_std_2</td>
                <td>–0.08</td>
                <td>.002</td>
                <td>—</td>
                <td>—</td>
                <td colspan="2">–0.09</td>
                <td>.005</td>
              </tr>
              <tr valign="top">
                <td>mfcc_std_3</td>
                <td>–0.07</td>
                <td>.002</td>
                <td>–0.10</td>
                <td>.002</td>
                <td colspan="2">—</td>
                <td>—</td>
              </tr>
              <tr valign="top">
                <td>focusfuture</td>
                <td>–0.07</td>
                <td>.003</td>
                <td>—</td>
                <td>—</td>
                <td colspan="2">–0.08</td>
                <td>.02</td>
              </tr>
              <tr valign="top">
                <td>mfcc_mean_2</td>
                <td>–0.07</td>
                <td>.004</td>
                <td>–0.09</td>
                <td>.01</td>
                <td colspan="2">—</td>
                <td>—</td>
              </tr>
              <tr valign="top">
                <td>adverb</td>
                <td>–0.07</td>
                <td>.004</td>
                <td>–0.11</td>
                <td>&#60;.001</td>
                <td colspan="2">—</td>
                <td>—</td>
              </tr>
              <tr valign="top">
                <td>time</td>
                <td>–0.07</td>
                <td>.004</td>
                <td>—</td>
                <td>—</td>
                <td colspan="2">–0.10</td>
                <td>.004</td>
              </tr>
              <tr valign="top">
                <td>function</td>
                <td>–0.07</td>
                <td>.005</td>
                <td>—</td>
                <td>—</td>
                <td colspan="2">—</td>
                <td>—</td>
              </tr>
              <tr valign="top">
                <td>negate</td>
                <td>0.07</td>
                <td>.006</td>
                <td>—</td>
                <td>—</td>
                <td colspan="2">0.08</td>
                <td>.01</td>
              </tr>
              <tr valign="top">
                <td>prep</td>
                <td>–0.06</td>
                <td>.007</td>
                <td>—</td>
                <td>—</td>
                <td colspan="2">–0.08</td>
                <td>.02</td>
              </tr>
              <tr valign="top">
                <td>WPS</td>
                <td>–0.06</td>
                <td>.007</td>
                <td>–0.07</td>
                <td>.03</td>
                <td colspan="2">—</td>
                <td>—</td>
              </tr>
              <tr valign="top">
                <td>anx</td>
                <td>0.06</td>
                <td>.008</td>
                <td>—</td>
                <td>—</td>
                <td colspan="2">0.08</td>
                <td>.01</td>
              </tr>
              <tr valign="top">
                <td>f0_std</td>
                <td>0.06</td>
                <td>.01</td>
                <td>—</td>
                <td>—</td>
                <td colspan="2">0.07</td>
                <td>.04</td>
              </tr>
              <tr valign="top">
                <td>hear</td>
                <td>0.06</td>
                <td>.01</td>
                <td>—</td>
                <td>—</td>
                <td colspan="2">0.10</td>
                <td>.003</td>
              </tr>
              <tr valign="top">
                <td>mfcc_std_5</td>
                <td>–0.06</td>
                <td>.01</td>
                <td>—</td>
                <td>—</td>
                <td colspan="2">–0.09</td>
                <td>.01</td>
              </tr>
              <tr valign="top">
                <td>death</td>
                <td>0.06</td>
                <td>.01</td>
                <td>0.07</td>
                <td>.04</td>
                <td colspan="2">—</td>
                <td>—</td>
              </tr>
              <tr valign="top">
                <td>ipron</td>
                <td>–0.06</td>
                <td>.01</td>
                <td>—</td>
                <td>—</td>
                <td colspan="2">–0.07</td>
                <td>.04</td>
              </tr>
              <tr valign="top">
                <td>see</td>
                <td>–0.06</td>
                <td>.01</td>
                <td>-0.09</td>
                <td>.006</td>
                <td colspan="2">—</td>
                <td>—</td>
              </tr>
              <tr valign="top">
                <td>affect</td>
                <td>0.06</td>
                <td>.02</td>
                <td>—</td>
                <td>—</td>
                <td colspan="2">0.07</td>
                <td>.04</td>
              </tr>
              <tr valign="top">
                <td>i</td>
                <td>0.05</td>
                <td>.02</td>
                <td>—</td>
                <td>—</td>
                <td colspan="2">—</td>
                <td>—</td>
              </tr>
              <tr valign="top">
                <td>family</td>
                <td>0.05</td>
                <td>.02</td>
                <td>—</td>
                <td>—</td>
                <td colspan="2">0.08</td>
                <td>.02</td>
              </tr>
              <tr valign="top">
                <td>mfcc_std_4</td>
                <td>–0.05</td>
                <td>.03</td>
                <td>—</td>
                <td>—</td>
                <td colspan="2">–0.07</td>
                <td>.04</td>
              </tr>
              <tr valign="top">
                <td>sad</td>
                <td>0.05</td>
                <td>.03</td>
                <td>0.08</td>
                <td>.01</td>
                <td colspan="2">—</td>
                <td>—</td>
              </tr>
              <tr valign="top">
                <td>ppron</td>
                <td>0.05</td>
                <td>.03</td>
                <td>—</td>
                <td>—</td>
                <td colspan="2">0.09</td>
                <td>.01</td>
              </tr>
              <tr valign="top">
                <td>space</td>
                <td>–0.05</td>
                <td>.04</td>
                <td>—</td>
                <td>—</td>
                <td colspan="2">—</td>
                <td>—</td>
              </tr>
              <tr valign="top">
                <td>article</td>
                <td>–0.05</td>
                <td>.04</td>
                <td>—</td>
                <td>—</td>
                <td colspan="2">–0.08</td>
                <td>.01</td>
              </tr>
              <tr valign="top">
                <td>leisure</td>
                <td>0.05</td>
                <td>.04</td>
                <td>—</td>
                <td>—</td>
                <td colspan="2">0.10</td>
                <td>.002</td>
              </tr>
              <tr valign="top">
                <td>friend</td>
                <td>0.05</td>
                <td>.047</td>
                <td>—</td>
                <td>—</td>
                <td colspan="2">—</td>
                <td>—</td>
              </tr>
              <tr valign="top">
                <td>lpcc_std_6</td>
                <td>—</td>
                <td>—</td>
                <td>–0.09</td>
                <td>.008</td>
                <td colspan="2">—</td>
                <td>—</td>
              </tr>
              <tr valign="top">
                <td>lpcc_std_4</td>
                <td>—</td>
                <td>—</td>
                <td>–0.09</td>
                <td>.008</td>
                <td colspan="2">—</td>
                <td>—</td>
              </tr>
              <tr valign="top">
                <td>intensity_mean</td>
                <td>—</td>
                <td>—</td>
                <td>–0.09</td>
                <td>.01</td>
                <td colspan="2">—</td>
                <td>—</td>
              </tr>
              <tr valign="top">
                <td>mfcc_mean_1</td>
                <td>—</td>
                <td>—</td>
                <td>–0.09</td>
                <td>.01</td>
                <td colspan="2">—</td>
                <td>—</td>
              </tr>
              <tr valign="top">
                <td>Dic</td>
                <td>—</td>
                <td>—</td>
                <td>–0.08</td>
                <td>.02</td>
                <td colspan="2">—</td>
                <td>—</td>
              </tr>
              <tr valign="top">
                <td>power</td>
                <td>—</td>
                <td>—</td>
                <td>0.07</td>
                <td>.03</td>
                <td colspan="2">–0.09</td>
                <td>.01</td>
              </tr>
              <tr valign="top">
                <td>lpcc_std_10</td>
                <td>—</td>
                <td>—</td>
                <td>–0.07</td>
                <td>.03</td>
                <td colspan="2">—</td>
                <td>—</td>
              </tr>
              <tr valign="top">
                <td>intensity_std</td>
                <td>—</td>
                <td>—</td>
                <td>–0.07</td>
                <td>.03</td>
                <td colspan="2">—</td>
                <td>—</td>
              </tr>
              <tr valign="top">
                <td>lpcc_std_12</td>
                <td>—</td>
                <td>—</td>
                <td>–0.07</td>
                <td>.04</td>
                <td colspan="2">—</td>
                <td>—</td>
              </tr>
              <tr valign="top">
                <td>mfcc_mean_8</td>
                <td>—</td>
                <td>—</td>
                <td>0.07</td>
                <td>.04</td>
                <td colspan="2">—</td>
                <td>—</td>
              </tr>
              <tr valign="top">
                <td>percept</td>
                <td>—</td>
                <td>—</td>
                <td>–0.07</td>
                <td>.046</td>
                <td colspan="2">—</td>
                <td>—</td>
              </tr>
              <tr valign="top">
                <td>lpcc_mean_4</td>
                <td>—</td>
                <td>—</td>
                <td>0.07</td>
                <td>.049</td>
                <td colspan="2">—</td>
                <td>—</td>
              </tr>
              <tr valign="top">
                <td>Apostro</td>
                <td>—</td>
                <td>—</td>
                <td>—</td>
                <td>—</td>
                <td colspan="2">0.09</td>
                <td>.005</td>
              </tr>
              <tr valign="top">
                <td>Sixltr</td>
                <td>—</td>
                <td>—</td>
                <td>—</td>
                <td>—</td>
                <td colspan="2">–0.09</td>
                <td>.01</td>
              </tr>
              <tr valign="top">
                <td>mfcc_mean_5</td>
                <td>—</td>
                <td>—</td>
                <td>—</td>
                <td>—</td>
                <td colspan="2">–0.08</td>
                <td>.02</td>
              </tr>
              <tr valign="top">
                <td>mfcc_std_11</td>
                <td>—</td>
                <td>—</td>
                <td>—</td>
                <td>—</td>
                <td colspan="2">–0.07</td>
                <td>.045</td>
              </tr>
              <tr valign="top">
                <td>f1_mean</td>
                <td>—</td>
                <td>—</td>
                <td>—</td>
                <td>—</td>
                <td colspan="2">0.07</td>
                <td> .047</td>
              </tr>
            </tbody>
          </table>
          <table-wrap-foot>
            <fn id="table1fn1">
              <p><sup>a</sup>Not available because the correlation was not significant.</p>
            </fn>
          </table-wrap-foot>
        </table-wrap>
      </sec>
      <sec>
        <title>Construction and Evaluation of Classification Models</title>
        <p>In this study, we aimed to evaluate the performance of a binary classifier that predicts if a person’s speech sample is in the “anxious” or “nonanxious” class based on the features of speech. The binary classification label is determined by processing the GAD-7 scale (which ranges from 0 to 21 in value) into 2 classes, anxious (GAD-7≥10) and nonanxious (GAD-7&#60;10), where 10 is a well-established screening threshold [<xref ref-type="bibr" rid="ref12">12</xref>] for GAD.</p>
        <p>An LR model was trained on the training data to make predictions between the 2 classes. The construction and evaluation steps were as follows. First, the input features were normalized, so that each feature would have a mean of 0 and an SD of 1. Next, the data were undersampled to equalize the representations from both the anxious and nonanxious classes. This avoided the problem of class imbalance, which, if occurred, caused low predictive accuracy for the minority class (which was the anxious class in our case). Therefore, samples were randomly selected and removed from the majority class until the majority class had an equal number of samples to the minority class.</p>
        <p>The model construction and training step used 3 data sets: a training data set, which was used to train the model; a validation data set, which was used to select the best hyperparameters during training; and a test data set, which was used to evaluate the performance of the trained model using area under the receiver operating characteristic (AUROC) metrics. These data sets were created within each sampling of the cross-validation (CV) scheme described next.</p>
        <p>The CV scheme used a nested resampling with 2-level nested CVs—one CV nested within another [<xref ref-type="bibr" rid="ref38">38</xref>]. In the outer loop, the data were split into 20% test data and 80% training and validation data. In the inner loop, 80% of the training and validation data were further split into 20% validation data and 80% training data. The inner loop was repeated 5 times, each with a different sampling to obtain a different 20/80 split. For each such split, the best hyperparameters were selected to maximize the accuracy of the validation data after training on the training data of the inner loop. After selecting the best hyperparameters from the inner CV loop, training was once again performed on the entire 80% of the outer loop training plus validation data, and the mean AUROC results were reported on the test data of the outer loop. The outer loop was iterated 5 times, each time selecting a different 20% for test data, until all the samples were left out and tested. This whole process was repeated 7 times, each with different random undersampling seeds, where in each of the 7 iterations, 5 AUROC were reported from the outer CV loop, giving a total of 35 values. The mean and SD of the AUROC values were used as the final metrics in this study to measure performance.</p>
      </sec>
      <sec>
        <title>Feature Selection</title>
        <p>During the construction of the model, a subset of features was selected from the features listed in <xref ref-type="table" rid="table1">Table 1</xref>. The goal of feature selection was to avoid using duplicate information (where the same information was present in different features) and maximize the prediction performance of our model.</p>
        <p>To avoid the use of duplicate information, we first calculated the intercorrelations between all the features presented in <xref ref-type="table" rid="table1">Table 1</xref>. We then used only one of each pair of the highly correlated (<italic>r</italic>&#62;0.8) features.</p>
        <p>In the model construction, it might not always be true that using all the available features maximizes the prediction accuracy; doing so may actually reduce accuracy owing to overfitting [<xref ref-type="bibr" rid="ref39">39</xref>]. Thus, to maximize the prediction performance of our model, we selected a subset of features using the following method: we began with the single feature that had the highest correlation with the GAD-7 and then measured the prediction performance of a trained model (on a validation data) using only that feature. Subsequent features were then added one-by-one in order of correlation until all the significant features (presented in <xref ref-type="table" rid="table1">Table 1</xref>) were used (ie, until adding 1 more feature no longer improved the prediction performance).</p>
      </sec>
      <sec>
        <title>Statistical Analysis</title>
        <p>To evaluate the performance of the prediction models, the mean AUROC of the 35 models was compared with the mean AUROC of a model that made a random prediction (ie, mean AUROC close to 0.5) using a modified 1-tailed <italic>t</italic> test developed by Bouckaert and Frank [<xref ref-type="bibr" rid="ref40">40</xref>]. The modified <italic>t</italic> test considers the fact that the individual AUROC values are not independent from each other, whereas in the original <italic>t</italic> test, the samples are expected to be independent. In our case, because the AUROC generated from a model shared some training data (owing to multiple undersampling and the 5-fold CV) with another, the AUROC values were not independent of each other. In our results, we considered a statistically significant difference at a <italic>P</italic> value significance level of .05.</p>
      </sec>
      <sec>
        <title>Ethics Approval</title>
        <p>This study was approved by the University of Toronto Research Ethics Board (protocol #37584).</p>
      </sec>
    </sec>
    <sec sec-type="results">
      <title>Results</title>
      <sec>
        <title>Recruitment and Data Inclusion</title>
        <p>A total of 2000 participants provided acceptable submissions from November 23, 2020, to May 28, 2021, and thus received payments. We reviewed the input data and audio for quality and included 1744 participants in the analysis. A detailed description of recruitment and data quality filtering was provided in our previous study [<xref ref-type="bibr" rid="ref11">11</xref>].</p>
      </sec>
      <sec>
        <title>Data Overview and Demographics of Participants</title>
        <p>Of the 1744 participants, 540 (30.96%) were above the GAD-7 screening threshold of 10 and 1204 (69.04%) were below the GAD-7 screening threshold of 10. Hereon, we will refer to those participants with a GAD-7 score ≥10 as the <italic>anxious class</italic> and those with a GAD-7 score &#60;10 as the <italic>nonanxious class</italic>.</p>
        <p><xref ref-type="table" rid="table2">Table 2</xref> shows participant demographics obtained from the Prolific recruitment platform. Column 1 of <xref ref-type="table" rid="table2">Table 2</xref> provide the names of demographic attributes and each category, while columns 2 and 3 give the number (and percentage) of participants with that attribute in the anxious and nonanxious groups, respectively. The last column gives the <italic>P</italic> values for the chi-square test of the null hypothesis that the difference in categories is independent, to determine if there is a significant difference between the anxious and nonanxious groups, for each categorical factor.</p>
        <table-wrap position="float" id="table2">
          <label>Table 2</label>
          <caption>
            <p>Demographics split by anxious and nonanxious label and chi-square test (N=1744).</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="30"/>
            <col width="270"/>
            <col width="0"/>
            <col width="250"/>
            <col width="0"/>
            <col width="250"/>
            <col width="0"/>
            <col width="0"/>
            <col width="200"/>
            <thead>
              <tr valign="top">
                <td colspan="3">Demographic factors</td>
                <td colspan="2">Anxious, n (%) (n=540)</td>
                <td colspan="2">Nonanxious, n (%) (n=1204)</td>
                <td colspan="2">Chi-square test, <italic>P</italic> value</td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td colspan="8">
                  <bold>Sex</bold>
                </td>
                <td>&#60;.001</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>Male</td>
                <td colspan="2">229 (26)</td>
                <td colspan="2">653 (74)</td>
                <td colspan="3">
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>Female</td>
                <td colspan="2">311 (36.1)</td>
                <td colspan="2">551 (63.9)</td>
                <td colspan="3">
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td colspan="8">
                  <bold>Self-reported ongoing mental health illness or condition</bold>
                </td>
                <td>&#60;.001</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>Yes</td>
                <td colspan="2">297 (48.8)</td>
                <td colspan="2">311 (51.2)</td>
                <td colspan="3">
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>No</td>
                <td colspan="2">243 (21.4)</td>
                <td colspan="2">893 (78.6)</td>
                <td colspan="3">
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td colspan="8">
                  <bold>Personal income (£<sup>a</sup>)</bold>
                </td>
                <td>&#60;.001</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>&#60;10,000</td>
                <td colspan="2">181 (39.2)</td>
                <td colspan="2">281 (60.8)</td>
                <td colspan="3">
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>10,000-19,999</td>
                <td colspan="2">112 (35)</td>
                <td colspan="2">208 (65)</td>
                <td colspan="3">
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>20,000-29,999</td>
                <td colspan="2">92 (26.2)</td>
                <td colspan="2">259 (73.8)</td>
                <td colspan="3">
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>30,000-39,999</td>
                <td colspan="2">60 (24.6)</td>
                <td colspan="2">184 (75.4)</td>
                <td colspan="3">
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>40,000-49,999</td>
                <td colspan="2">36 (24.8)</td>
                <td colspan="2">109 (75.2)</td>
                <td colspan="3">
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>50,000-59,999</td>
                <td colspan="2">20 (21.3)</td>
                <td colspan="2">74 (78.7)</td>
                <td colspan="3">
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>≥60,000</td>
                <td colspan="2">39 (30.5)</td>
                <td colspan="2">89 (69.5)</td>
                <td colspan="3">
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td colspan="8">
                  <bold>Age (years)</bold>
                </td>
                <td>&#60;.001</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>18-19</td>
                <td colspan="2">27 (38)</td>
                <td colspan="2">44 (62)</td>
                <td colspan="3">
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>20-29</td>
                <td colspan="2">239 (38.7)</td>
                <td colspan="2">379 (61.3)</td>
                <td colspan="3">
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>30-39</td>
                <td colspan="2">162 (32.7)</td>
                <td colspan="2">334 (67.3)</td>
                <td colspan="3">
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>40-49</td>
                <td colspan="2">67 (23.4)</td>
                <td colspan="2">219 (76.6)</td>
                <td colspan="3">
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>50-59</td>
                <td colspan="2">39 (22.8)</td>
                <td colspan="2">132 (77.2)</td>
                <td colspan="3">
                  <break/>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>≥60</td>
                <td colspan="2">6 (5.9)</td>
                <td colspan="2">96 (94.1)</td>
                <td colspan="3">
                  <break/>
                </td>
              </tr>
            </tbody>
          </table>
          <table-wrap-foot>
            <fn id="table2fn1">
              <p><sup>a</sup>1 £=US $1.37.</p>
            </fn>
          </table-wrap-foot>
        </table-wrap>
      </sec>
      <sec>
        <title>Classification Model Performance</title>
        <p>In this section, the mean AUROC of a binary classification model that classified between anxious and nonanxious classes is presented. The following subsections summarize our main empirical results for different types of inputs to the classification models.</p>
        <sec>
          <title>Acoustic and Linguistic Features as Input</title>
          <p>The mean AUROC for the model constructed using a subset of the acoustic and linguistic features selected using the feature selection method described in the Methods section is reported. As described in the Methods section, the features that contain very similar information were not used based on the intercorrelation between the features. <xref ref-type="table" rid="table3">Table 3</xref> presents the features with high intercorrelation (<italic>r</italic>&#62;0.8) between the features presented in <xref ref-type="table" rid="table1">Table 1</xref>. We only considered using one of each pair of the highly intercorrelated features.</p>
          <table-wrap position="float" id="table3">
            <label>Table 3</label>
            <caption>
              <p>Features with high intercorrelation (similar features) with each other.</p>
            </caption>
            <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
              <col width="30"/>
              <col width="670"/>
              <col width="0"/>
              <col width="150"/>
              <col width="0"/>
              <col width="150"/>
              <thead>
                <tr valign="top">
                  <td colspan="3">Sample and feature</td>
                  <td colspan="2">
                    <italic>r</italic>
                  </td>
                  <td><italic>P</italic> value</td>
                </tr>
              </thead>
              <tbody>
                <tr valign="top">
                  <td colspan="6">
                    <bold>All samples</bold>
                  </td>
                </tr>
                <tr valign="top">
                  <td>
                    <break/>
                  </td>
                  <td>AllPunc, Period</td>
                  <td colspan="2">0.93</td>
                  <td colspan="2">&#60;.001</td>
                </tr>
                <tr valign="top">
                  <td>
                    <break/>
                  </td>
                  <td>i, ppron</td>
                  <td colspan="2">0.81</td>
                  <td colspan="2">&#60;.001</td>
                </tr>
                <tr valign="top">
                  <td colspan="6">
                    <bold>Female samples</bold>
                  </td>
                </tr>
                <tr valign="top">
                  <td>
                    <break/>
                  </td>
                  <td>AllPunc, Period</td>
                  <td colspan="2">0.93</td>
                  <td colspan="2">&#60;.001</td>
                </tr>
                <tr valign="top">
                  <td>
                    <break/>
                  </td>
                  <td>Intensity_mean, intensity_std</td>
                  <td colspan="2">0.93</td>
                  <td colspan="2">&#60;.001</td>
                </tr>
                <tr valign="top">
                  <td colspan="6">
                    <bold>Male samples</bold>
                  </td>
                </tr>
                <tr valign="top">
                  <td>
                    <break/>
                  </td>
                  <td>AllPunc, Period</td>
                  <td colspan="2">0.93</td>
                  <td colspan="2">&#60;.001</td>
                </tr>
              </tbody>
            </table>
          </table-wrap>
          <p>As described in the <italic>Methods</italic> section, the acoustic and linguistic features starting with the feature with the highest correlation were included in the model, incrementally, if they showed improvement in the performance of the model. <xref ref-type="table" rid="table4">Table 4</xref> shows the subset of features used for the 3 data sets, and <xref rid="figure1" ref-type="fig">Figure 1</xref> shows the mean AUROC as a function of the number of selected features. Using the feature selection method discussed in the Methods section, the number of features required to produce the maximum mean AUROC was 11 for the all-sample, 7 for the female sample, and 11 for the male sample data set, as shown in <xref ref-type="table" rid="table4">Table 4</xref>. The best model is the one that included all the features listed in <xref ref-type="table" rid="table4">Table 4</xref> (according to the data set).</p>
          <p><xref ref-type="table" rid="table5">Table 5</xref> shows the mean AUROC across the 35 data splits, as described in the Methods section. It also provides results of the 1-tailed <italic>t</italic> test comparison of the best model with that of a random model.</p>
          <table-wrap position="float" id="table4">
            <label>Table 4</label>
            <caption>
              <p>Subset of acoustic and linguist features used in the 3 models after feature selection.</p>
            </caption>
            <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
              <col width="280"/>
              <col width="120"/>
              <col width="120"/>
              <col width="120"/>
              <col width="120"/>
              <col width="120"/>
              <col width="120"/>
              <thead>
                <tr valign="top">
                  <td>Feature</td>
                  <td colspan="6">Data set</td>
                </tr>
                <tr valign="top">
                  <td>
                    <break/>
                  </td>
                  <td colspan="2">All-sample</td>
                  <td colspan="2">Female sample</td>
                  <td colspan="2">Male sample</td>
                </tr>
                <tr valign="top">
                  <td>
                    <break/>
                  </td>
                  <td>
                    <italic>r</italic>
                  </td>
                  <td><italic>P</italic> value</td>
                  <td>
                    <italic>r</italic>
                  </td>
                  <td><italic>P</italic> value</td>
                  <td>
                    <italic>r</italic>
                  </td>
                  <td><italic>P</italic> value</td>
                </tr>
              </thead>
              <tbody>
                <tr valign="top">
                  <td>AllPunc</td>
                  <td>0.13</td>
                  <td>&#60;.001</td>
                  <td>—<sup>a</sup></td>
                  <td>—</td>
                  <td>—</td>
                  <td>—</td>
                </tr>
                <tr valign="top">
                  <td>Assent</td>
                  <td>0.1</td>
                  <td>&#60;.001</td>
                  <td>—</td>
                  <td>—</td>
                  <td>0.11</td>
                  <td>.001</td>
                </tr>
                <tr valign="top">
                  <td>Relativ</td>
                  <td>−0.09</td>
                  <td>&#60;.001</td>
                  <td>—</td>
                  <td>
                    <break/>
                  </td>
                  <td>−0.1</td>
                  <td>.002</td>
                </tr>
                <tr valign="top">
                  <td>Motion</td>
                  <td>−0.08</td>
                  <td>&#60;.001</td>
                  <td>−0.1</td>
                  <td>.003</td>
                  <td>—</td>
                  <td>—</td>
                </tr>
                <tr valign="top">
                  <td>mfcc_std_2</td>
                  <td>−0.08</td>
                  <td>.002</td>
                  <td>—</td>
                  <td>—</td>
                  <td>—</td>
                  <td>—</td>
                </tr>
                <tr valign="top">
                  <td>mfcc_std_3</td>
                  <td>−0.07</td>
                  <td>.002</td>
                  <td>—</td>
                  <td>—</td>
                  <td>—</td>
                  <td>—</td>
                </tr>
                <tr valign="top">
                  <td>Focusfuture</td>
                  <td>−0.07</td>
                  <td>.003</td>
                  <td>—</td>
                  <td>—</td>
                  <td>−0.08</td>
                  <td>.02</td>
                </tr>
                <tr valign="top">
                  <td>mfcc_std_5</td>
                  <td>−0.06</td>
                  <td>.01</td>
                  <td>—</td>
                  <td>—</td>
                  <td>—</td>
                  <td>—</td>
                </tr>
                <tr valign="top">
                  <td>Death</td>
                  <td>0.06</td>
                  <td>.01</td>
                  <td>0.07</td>
                  <td>.04</td>
                  <td>—</td>
                  <td>—</td>
                </tr>
                <tr valign="top">
                  <td>See</td>
                  <td>−0.06</td>
                  <td>.01</td>
                  <td>−0.09</td>
                  <td>.006</td>
                  <td>—</td>
                  <td>—</td>
                </tr>
                <tr valign="top">
                  <td>mfcc_std_4</td>
                  <td>−0.05</td>
                  <td>.045</td>
                  <td>—</td>
                  <td>—</td>
                  <td>—</td>
                  <td>—</td>
                </tr>
                <tr valign="top">
                  <td>Period</td>
                  <td>—</td>
                  <td>—</td>
                  <td>0.16</td>
                  <td>&#60;.001</td>
                  <td>—</td>
                  <td>—</td>
                </tr>
                <tr valign="top">
                  <td>Dic</td>
                  <td>—</td>
                  <td>—</td>
                  <td>−0.08</td>
                  <td>.02</td>
                  <td>—</td>
                  <td>—</td>
                </tr>
                <tr valign="top">
                  <td>Power</td>
                  <td>—</td>
                  <td>—</td>
                  <td>0.07</td>
                  <td>.03</td>
                  <td>—</td>
                  <td>—</td>
                </tr>
                <tr valign="top">
                  <td>lpcc_std_10</td>
                  <td>—</td>
                  <td>—</td>
                  <td>−0.07</td>
                  <td>.03</td>
                  <td>—</td>
                  <td>—</td>
                </tr>
                <tr valign="top">
                  <td>speaking_duration</td>
                  <td>—</td>
                  <td>—</td>
                  <td>—</td>
                  <td>—</td>
                  <td>−0.13</td>
                  <td>&#60;.001</td>
                </tr>
                <tr valign="top">
                  <td>Leisure</td>
                  <td>—</td>
                  <td>—</td>
                  <td>—</td>
                  <td>—</td>
                  <td>0.1</td>
                  <td>.002</td>
                </tr>
                <tr valign="top">
                  <td>Time</td>
                  <td>—</td>
                  <td>—</td>
                  <td>—</td>
                  <td>—</td>
                  <td>−0.1</td>
                  <td>.004</td>
                </tr>
                <tr valign="top">
                  <td>Ppron</td>
                  <td>—</td>
                  <td>—</td>
                  <td>—</td>
                  <td>—</td>
                  <td>0.09</td>
                  <td>.01</td>
                </tr>
                <tr valign="top">
                  <td>Negemo</td>
                  <td>—</td>
                  <td>—</td>
                  <td>—</td>
                  <td>—</td>
                  <td>0.08</td>
                  <td>.01</td>
                </tr>
                <tr valign="top">
                  <td>Article</td>
                  <td>—</td>
                  <td>—</td>
                  <td>—</td>
                  <td>—</td>
                  <td>−0.08</td>
                  <td>.01</td>
                </tr>
                <tr valign="top">
                  <td>mfcc_mean_5</td>
                  <td>—</td>
                  <td>—</td>
                  <td>—</td>
                  <td>—</td>
                  <td>−0.08</td>
                  <td>.01</td>
                </tr>
                <tr valign="top">
                  <td>f1_mean</td>
                  <td>—</td>
                  <td>—</td>
                  <td>—</td>
                  <td>—</td>
                  <td>0.07</td>
                  <td>.047</td>
                </tr>
              </tbody>
            </table>
            <table-wrap-foot>
              <fn id="table4fn1">
                <p><sup>a</sup>Not available because the correlation was not significant.</p>
              </fn>
            </table-wrap-foot>
          </table-wrap>
          <fig id="figure1" position="float">
            <label>Figure 1</label>
            <caption>
              <p>Mean area under the receiver operating characteristic (AUROC) as a function of the number of selected features.</p>
            </caption>
            <graphic xlink:href="formative_v6i10e39998_fig1.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
          </fig>
          <table-wrap position="float" id="table5">
            <label>Table 5</label>
            <caption>
              <p>Mean area under the receiver operating characteristic (AUROC) of a model trained using the subset of features (N=1744).</p>
            </caption>
            <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
              <col width="340"/>
              <col width="330"/>
              <col width="170"/>
              <col width="160"/>
              <thead>
                <tr valign="top">
                  <td>Data set</td>
                  <td>AUROC, mean (SD)</td>
                  <td><italic>t</italic> test (<italic>df</italic>)</td>
                  <td><italic>P</italic> value</td>
                </tr>
              </thead>
              <tbody>
                <tr valign="top">
                  <td>All-sample</td>
                  <td>0.59 (0.02)</td>
                  <td>4.93 (34)</td>
                  <td>&#60;.001</td>
                </tr>
                <tr valign="top">
                  <td>Female sample (n=862)</td>
                  <td>0.60 (0.04)</td>
                  <td>4.25 (34)</td>
                  <td>&#60;.001</td>
                </tr>
                <tr valign="top">
                  <td>Male sample (n=882)</td>
                  <td>0.61 (0.06)</td>
                  <td>4.21 (34)</td>
                  <td>&#60;.001</td>
                </tr>
              </tbody>
            </table>
          </table-wrap>
        </sec>
        <sec>
          <title>Using Participants’ Demographics</title>
          <p>This section presents the performance of the model when augmented with age, sex, and income demographic information. <xref ref-type="table" rid="table6">Table 6</xref> shows the mean AUROC of the LR model that used both demographic information and acoustic and linguistic features. It also included a modified <italic>t</italic> test comparison with a random model.</p>
          <p><xref ref-type="table" rid="table7">Table 7</xref> shows the results of the <italic>t</italic> test between the model with only acoustic or linguistic features and the model that also used demographic information. <xref ref-type="table" rid="table8">Table 8</xref> separates each of the demographic features and shows the mean AUROC of these models when using a single demographic at a time, together with the acoustic and linguistic features.</p>
          <table-wrap position="float" id="table6">
            <label>Table 6</label>
            <caption>
              <p>Mean area under the receiver operating characteristic (AUROC) of a model trained using demographic information (age, sex, and income) in addition to the acoustic and linguistic features and comparison with a random model (N=1744).</p>
            </caption>
            <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
              <col width="340"/>
              <col width="330"/>
              <col width="170"/>
              <col width="160"/>
              <thead>
                <tr valign="top">
                  <td>Data set</td>
                  <td>AUROC, mean (SD)</td>
                  <td><italic>t</italic> test (<italic>df</italic>)</td>
                  <td><italic>P</italic> value</td>
                </tr>
              </thead>
              <tbody>
                <tr valign="top">
                  <td>All-sample</td>
                  <td>0.64 (0.03)</td>
                  <td>6.21 (34)</td>
                  <td>&#60;.001</td>
                </tr>
                <tr valign="top">
                  <td>Female sample (n=862)</td>
                  <td>0.66 (0.04)</td>
                  <td>5.89 (34)</td>
                  <td>&#60;.001</td>
                </tr>
                <tr valign="top">
                  <td>Male sample (n=882)</td>
                  <td>0.62 (0.07)</td>
                  <td>4.36 (34)</td>
                  <td>&#60;.001</td>
                </tr>
              </tbody>
            </table>
          </table-wrap>
          <table-wrap position="float" id="table7">
            <label>Table 7</label>
            <caption>
              <p>Comparison of model trained using only acoustic or linguistic features with model that also uses demographic information (N=1744).</p>
            </caption>
            <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
              <col width="200"/>
              <col width="280"/>
              <col width="320"/>
              <col width="0"/>
              <col width="100"/>
              <col width="0"/>
              <col width="100"/>
              <thead>
                <tr valign="top">
                  <td>Data set</td>
                  <td colspan="3">AUROC<sup>a</sup>, mean (SD)</td>
                  <td colspan="2"><italic>t</italic> test (<italic>df</italic>)</td>
                  <td><italic>P</italic> value</td>
                </tr>
                <tr valign="top">
                  <td>
                    <break/>
                  </td>
                  <td>Acoustic and linguistic features</td>
                  <td>Demographics, acoustic and linguistic features</td>
                  <td colspan="2">
                    <break/>
                  </td>
                  <td colspan="2">
                    <break/>
                  </td>
                </tr>
              </thead>
              <tbody>
                <tr valign="top">
                  <td>All-sample</td>
                  <td>0.59 (0.02)</td>
                  <td>0.64 (0.03)</td>
                  <td colspan="2">4.01 (34)</td>
                  <td colspan="2">&#60;.001</td>
                </tr>
                <tr valign="top">
                  <td>Female sample (n=862)</td>
                  <td>0.60 (0.04)</td>
                  <td>0.66 (0.04)</td>
                  <td colspan="2">4.21 (34)</td>
                  <td colspan="2">&#60;.001</td>
                </tr>
                <tr valign="top">
                  <td>Male sample (n=882)</td>
                  <td>0.61 (0.06)</td>
                  <td>0.62 (0.07)</td>
                  <td colspan="2">0.76 (34)</td>
                  <td colspan="2">.45</td>
                </tr>
              </tbody>
            </table>
            <table-wrap-foot>
              <fn id="table7fn1">
                <p><sup>a</sup>AUROC: area under the receiver operating characteristic.</p>
              </fn>
            </table-wrap-foot>
          </table-wrap>
          <table-wrap position="float" id="table8">
            <label>Table 8</label>
            <caption>
              <p>Mean area under the receiver operating characteristic (AUROC) of the model when adding a single demographic characteristic to the acoustic and linguistic features (N=1744).</p>
            </caption>
            <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
              <col width="250"/>
              <col width="250"/>
              <col width="250"/>
              <col width="250"/>
              <thead>
                <tr valign="top">
                  <td>Data set</td>
                  <td colspan="3">AUROC of model using acoustic and linguistic features and including o<italic>nly one of the demographics,</italic> mean (SD)</td>
                </tr>
                <tr valign="top">
                  <td>
                    <break/>
                  </td>
                  <td>Age</td>
                  <td>Income</td>
                  <td>Sex</td>
                </tr>
              </thead>
              <tbody>
                <tr valign="top">
                  <td>All-sample</td>
                  <td>0.64 (0.03)</td>
                  <td>0.6 (0.02)</td>
                  <td>0.59 (0.02)</td>
                </tr>
                <tr valign="top">
                  <td>Female sample (n=862)</td>
                  <td>0.66 (0.04)</td>
                  <td>0.6 (0.05)</td>
                  <td>N/A<sup>a</sup></td>
                </tr>
                <tr valign="top">
                  <td>Male sample (n=882)</td>
                  <td>0.62 (0.07)</td>
                  <td>0.61 (0.06)</td>
                  <td>N/A</td>
                </tr>
              </tbody>
            </table>
            <table-wrap-foot>
              <fn id="table8fn1">
                <p><sup>a</sup>N/A: not applicable.</p>
              </fn>
            </table-wrap-foot>
          </table-wrap>
        </sec>
      </sec>
    </sec>
    <sec sec-type="discussion">
      <title>Discussion</title>
      <p>The main objective of this study was to investigate the prediction performance of a model that screens for GAD from acoustic and linguistic features of impromptu speech. To do so, we have explored an LR model, and in the following subsections, we discuss the findings presented in the <italic>Results</italic> section, as well as the limitations of this study.</p>
      <sec>
        <title>Principal Findings</title>
        <sec>
          <title>Recruitment and Data Inclusion</title>
          <p>As the study continued from November 23, 2020, to May 28, 2021, the recruitment took place during the global COVID-19 pandemic. We speculated that this might have resulted in an above-normal number of participants who work remotely using their personal computers, hence making web-based recruitments relatively quicker.</p>
        </sec>
        <sec>
          <title>Demographics of Participants</title>
          <p>The percentage of anxious and nonanxious participants shows that the anxious group made up to 30.96% (540/1744) of the total, which is much higher than the general population rate of 10% [<xref ref-type="bibr" rid="ref2">2</xref>]. Previous studies [<xref ref-type="bibr" rid="ref11">11</xref>,<xref ref-type="bibr" rid="ref26">26</xref>-<xref ref-type="bibr" rid="ref28">28</xref>,<xref ref-type="bibr" rid="ref41">41</xref>] that also used participants recruited from Prolific showed a higher number of anxious participants in the recruitment pool. <xref ref-type="table" rid="table2">Table 2</xref> sheds some light on this difference, showing that a similarly high fraction of participants self-reported on their Prolific profile that they have an ongoing mental health condition.</p>
          <p>We also aimed to obtain broader demographics than those in the prior study. Most prior studies focused on a certain type of demographics, such as the study by McGinnis et al [<xref ref-type="bibr" rid="ref13">13</xref>], which focused on children, and the studies by Week et al [<xref ref-type="bibr" rid="ref17">17</xref>], Salekin et al [<xref ref-type="bibr" rid="ref20">20</xref>], and Rook et al [<xref ref-type="bibr" rid="ref23">23</xref>], which focused on undergraduate students. Both these types significantly limited the age range of the participants. The data presented in <xref ref-type="table" rid="table2">Table 2</xref> show that the age range of our participants had a broader distribution. The same is true for personal income, which showed a range of economic status in our participant pool.</p>
        </sec>
        <sec>
          <title>Acoustic and Linguistic Features as Input</title>
          <p>The LR used statistically significant acoustic and linguistic features selected by the feature selection method discussed in the <italic>Methods</italic> section and presented in <xref ref-type="table" rid="table4">Table 4</xref>. Although the correlation between the features used and the GAD-7 was very small (the highest being 0.13), the model built using these features was able to perform significantly better (with <italic>P</italic>&#60;.05) than a random model. The mean AUROC results presented in <xref ref-type="table" rid="table5">Table 5</xref> suggest that there is some signal to be detected from the combined effect of the acoustic and linguistic features of speech.</p>
          <p>Although it is possible to use the GAD-7 scale to screen for GAD (it has a sensitivity of 89% and a specificity of 82% [<xref ref-type="bibr" rid="ref12">12</xref>]), it cannot serve the purpose of our study, which is a continuous and passive monitoring of a participant. By contrast, an automated screener that listens passively to speech has the potential to frequently monitor speech samples from participants. Furthermore, the probability of correct prediction can be improved by using multiple measurements under the assumption that each measurement from different speech is relatively independent.</p>
          <p>This enhanced accuracy could be achieved by considering the model’s native accuracy as follows: let the accuracy of a correct prediction from a single measurement be <italic>a</italic>, and we take <italic>N</italic> successive measurements, based on <italic>N</italic> successive speech samples, using our model. As a decision procedure, we would decide that most of the measurement is correct—whichever result, anxious or nonanxious, happens in more than <italic>N/2</italic> of the measurements. We were interested in the probability that this decision procedure will produce a correct result. The probability that <italic>n</italic> or more of the <italic>N</italic> measurements would have a correct prediction can be calculated using the cumulative binomial distribution function (Inline graphic 1). Given the decision procedure of taking the correct result to be the majority result of the <italic>N</italic> trials, we set the value of n to be <italic>N/2</italic>, which computes the probability of more than <italic>N/2</italic> correct answers. As long as the single prediction <italic>a</italic> is &#62;0.5, the computed probability <italic>A</italic> will be &#62;<italic>a</italic>. It should be noted that this result does rely on the assumption that the measurements are independent when, in reality, they are not because the measurements were taken from the same person. However, the set of words coming from the person was different, and more spaced-out measurements might have reduced the dependency between the samples. To summarize, it is possible to increase the accuracy of correct predictions by taking multiple measurements and taking the class (anxious or nonanxious) that has been predicted most of the time as the final predicted value.</p>
          <graphic xlink:href="formative_v6i10e39998_fig2.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </sec>
        <sec>
          <title>Participants’ Demographics as Input Features</title>
          <p>In a scenario in which an anxiety screening or prediction model can be deployed, an individual’s demographic information can easily be collected. Thus, it is reasonable to explore the predictive capability of this additional information. <xref ref-type="table" rid="table6">Table 6</xref> shows that a model built using demographic information as input in addition to the acoustic and linguistic features was still able to perform significantly better than a random model. <xref ref-type="table" rid="table7">Table 7</xref> compares this model with a model that used only acoustic and linguistic features as input. The results show that the demographic information significantly improved the mean AUROC of the models built on the all-sample and female sample data sets but failed on the model built on the male sample data set.</p>
          <p>The impact of each demographic variable was explored separately. <xref ref-type="table" rid="table8">Table 8</xref> shows the mean AUROC of the model when only one of the demographics was used together with the acoustic and linguistic features. The addition of age affected the prediction performance of the model, whereas the addition of either sex or income did not show a significant improvement. In addition, the fact that the addition of age affected the prediction performance for the model built on female participants suggests that their anxiety depends on age compared with anxiety in male participants.</p>
          <p>Comparing our results with a prior study that aimed to predict GAD, there were studies that achieved the above-random prediction for GAD [<xref ref-type="bibr" rid="ref21">21</xref>,<xref ref-type="bibr" rid="ref23">23</xref>], and there were studies that did not [<xref ref-type="bibr" rid="ref28">28</xref>]. Our models performed significantly better than the random model, and we speculated that this might be attributed to the larger and demographically broad sample size that enabled our model to learn a large amount of information in predicting GAD. We also note that a prior study that did not succeed in predicting GAD [<xref ref-type="bibr" rid="ref28">28</xref>] did, in fact, succeed in predicting SAD. They believed that the symptoms of SAD might be more manifested in the participants’ behavior and, therefore speech, compared with GAD. Other studies focusing on SAD have also been successful in above-random prediction [<xref ref-type="bibr" rid="ref17">17</xref>,<xref ref-type="bibr" rid="ref20">20</xref>].</p>
        </sec>
      </sec>
      <sec>
        <title>Limitations</title>
        <p>A limitation of this study arises from the data collection method used with respect to the scenarios of use that were described in the Introduction section. We suggested that the prediction of anxiety from speech could be applied to passively collected speech data gathered while the patient is going through their daily activities. This could help in automated anxiety screening, treatment monitoring, and relapse detection. However, the data used in this study were actively collected when the participants spoke in front of a camera, and it may be substantially different from such passively collected speech. Future studies could investigate the models that we suggest using passively collected speech.</p>
        <p>Another limitation was the use of a web-based participant recruitment method. Individuals willing to work on a web-based participant recruitment platform may be limited to a particular type of demographics in a certain society. For example, we noted that in our recruitment pool, there was a higher percentage of anxious participants compared with the general population. In our study, we sought generalizability, and even though our participants were more diverse in terms of demographics compared with prior studies, it could be more generalizable if we recruited participants from sources other than the web-based recruitment.</p>
        <p>Another limitation was the artificial setup used to replicate the original TSST. In the original TSST, participants described their dream job in front of a live panel of judges. Owing to the restrictions that the COVID-19 pandemic had caused, we were not able to recruit participants for an in-person study; instead, we had participants describe their dream job in front of a camera at their own location (with different recording devices). Despite its limitations, this approach also had important benefits because it enabled us to recruit a large number of participants, which would otherwise have been extremely difficult for an in-person study.</p>
      </sec>
      <sec>
        <title>Conclusions</title>
        <p>In this study, we developed a model to predict the presence or absence of GAD based on the speech features. These speech features were chosen because prior studies have suggested that they are associated with other types of anxiety disorders including GAD. Our results have shown that it is possible to achieve the above-random prediction accuracy for GAD from the acoustic and linguistic features of speech while using a larger and more generalizable sample size. Prediction accuracy can also be further improved by adding basic demographic information. Even though we have investigated adding 3 different types of demographic variables (age, sex, and income), the most influential variable that showed improvement in prediction accuracy was age.</p>
        <p>Furthermore, we have discussed that the results from multiple measurements have the possibility to improve prediction accuracy. Therefore, we recommend that future studies explore the collection of multiple speech samples sampled throughout the day or week and investigate the extent to which the prediction accuracy can be improved. This will allow for the acoustic and linguistic features of speech, together with basic demographic information, to be used in a system to trigger early intervention, monitor treatment responses, or detect relapses.</p>
      </sec>
    </sec>
  </body>
  <back>
    <app-group>
      <supplementary-material id="app1">
        <label>Multimedia Appendix 1</label>
        <p>Description of the acoustic and linguistic features.</p>
        <media xlink:href="formative_v6i10e39998_app1.pdf" xlink:title="PDF File  (Adobe PDF File), 65 KB"/>
      </supplementary-material>
    </app-group>
    <glossary>
      <title>Abbreviations</title>
      <def-list>
        <def-item>
          <term id="abb1">AUROC</term>
          <def>
            <p>area under the receiver operating characteristic</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb2">BAI</term>
          <def>
            <p>Beck Anxiety Inventory</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb3">BIS/BAS</term>
          <def>
            <p>behavioral inhibition/behavioral approach scale</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb4">CV</term>
          <def>
            <p>cross-validation</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb5">F0</term>
          <def>
            <p>fundamental frequency</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb6">GAD</term>
          <def>
            <p>generalized anxiety disorder</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb7">GAD-7</term>
          <def>
            <p>Generalized Anxiety Disorder 7-item</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb8">LIWC</term>
          <def>
            <p>Linguistic Inquiry and Word Count</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb9">LR</term>
          <def>
            <p>logistic regression</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb10">SAD</term>
          <def>
            <p>social anxiety disorder</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb11">SVM</term>
          <def>
            <p>support vector machine</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb12">TSST</term>
          <def>
            <p>Trier Social Stress Test</p>
          </def>
        </def-item>
      </def-list>
    </glossary>
    <ack>
      <p>This research was funded by a University of Toronto XSeed grant, Natural Sciences and Engineering Research Council of Canada Discovery grant RGPIN-2019-04395, and Social Sciences and Humanities Research Council Partnership Engage grant 892-2019-0011.</p>
      <p>The authors are also grateful to Professor Ludovic Rheault for his energy and assistance with the launch of this project and the data collection and advice.</p>
    </ack>
    <fn-group>
      <fn fn-type="conflict">
        <p>None declared.</p>
      </fn>
    </fn-group>
    <ref-list>
      <ref id="ref1">
        <label>1</label>
        <nlm-citation citation-type="book">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Barlow</surname>
              <given-names>DH</given-names>
            </name>
          </person-group>
          <source>Anxiety and Its Disorders: The Nature and Treatment of Anxiety and Panic</source>
          <year>1988</year>
          <publisher-loc>New York, NY, USA</publisher-loc>
          <publisher-name>Guilford Press</publisher-name>
        </nlm-citation>
      </ref>
      <ref id="ref2">
        <label>2</label>
        <nlm-citation citation-type="web">
          <article-title>Mental Health - Anxiety Disorders</article-title>
          <source>Public Health Canada</source>
          <year>2009</year>
          <month>7</month>
          <access-date>2022-01-24</access-date>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.canada.ca/en/health-canada/services/healthy-living/your-health/diseases/mental-health-anxiety-disorders.html">https://www.canada.ca/en/health-canada/services/healthy-living/your-health/diseases/mental-health-anxiety-disorders.html</ext-link>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref3">
        <label>3</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Roberge</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Fournier</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Duhoux</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Nguyen</surname>
              <given-names>CT</given-names>
            </name>
            <name name-style="western">
              <surname>Smolders</surname>
              <given-names>M</given-names>
            </name>
          </person-group>
          <article-title>Mental health service use and treatment adequacy for anxiety disorders in Canada</article-title>
          <source>Soc Psychiatry Psychiatr Epidemiol</source>
          <year>2011</year>
          <month>04</month>
          <volume>46</volume>
          <issue>4</issue>
          <fpage>321</fpage>
          <lpage>30</lpage>
          <pub-id pub-id-type="doi">10.1007/s00127-010-0186-2</pub-id>
          <pub-id pub-id-type="medline">20217041</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref4">
        <label>4</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Koerner</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Dugas</surname>
              <given-names>MJ</given-names>
            </name>
            <name name-style="western">
              <surname>Savard</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Gaudet</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Turcotte</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Marchand</surname>
              <given-names>A</given-names>
            </name>
          </person-group>
          <article-title>The economic burden of anxiety disorders in Canada</article-title>
          <source>Can Psychol</source>
          <year>2004</year>
          <month>08</month>
          <volume>45</volume>
          <issue>3</issue>
          <fpage>191</fpage>
          <lpage>201</lpage>
          <pub-id pub-id-type="doi">10.1037/h0088236</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref5">
        <label>5</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Elliott</surname>
              <given-names>KP</given-names>
            </name>
            <name name-style="western">
              <surname>Hunsley</surname>
              <given-names>J</given-names>
            </name>
          </person-group>
          <article-title>Evaluating the measurement of mental health service accessibility, acceptability, and availability in the Canadian Community Health Survey</article-title>
          <source>Am J Orthopsychiatry</source>
          <year>2015</year>
          <month>05</month>
          <volume>85</volume>
          <issue>3</issue>
          <fpage>238</fpage>
          <lpage>42</lpage>
          <pub-id pub-id-type="doi">10.1037/ort0000062</pub-id>
          <pub-id pub-id-type="medline">25894307</pub-id>
          <pub-id pub-id-type="pii">2015-16492-001</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref6">
        <label>6</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Hidalgo</surname>
              <given-names>RB</given-names>
            </name>
            <name name-style="western">
              <surname>Sheehan</surname>
              <given-names>DV</given-names>
            </name>
          </person-group>
          <article-title>Generalized anxiety disorder</article-title>
          <source>Handb Clin Neurol</source>
          <year>2012</year>
          <volume>106</volume>
          <fpage>343</fpage>
          <lpage>62</lpage>
          <pub-id pub-id-type="doi">10.1016/B978-0-444-52002-9.00019-X</pub-id>
          <pub-id pub-id-type="medline">22608630</pub-id>
          <pub-id pub-id-type="pii">B978-0-444-52002-9.00019-X</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref7">
        <label>7</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Brown</surname>
              <given-names>TA</given-names>
            </name>
            <name name-style="western">
              <surname>Di Nardo</surname>
              <given-names>PA</given-names>
            </name>
            <name name-style="western">
              <surname>Lehman</surname>
              <given-names>CL</given-names>
            </name>
            <name name-style="western">
              <surname>Campbell</surname>
              <given-names>LA</given-names>
            </name>
          </person-group>
          <article-title>Reliability of DSM-IV anxiety and mood disorders: implications for the classification of emotional disorders</article-title>
          <source>J Abnorm Psychol</source>
          <year>2001</year>
          <month>02</month>
          <volume>110</volume>
          <issue>1</issue>
          <fpage>49</fpage>
          <lpage>58</lpage>
          <pub-id pub-id-type="doi">10.1037//0021-843x.110.1.49</pub-id>
          <pub-id pub-id-type="medline">11261399</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref8">
        <label>8</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Thompson</surname>
              <given-names>AR</given-names>
            </name>
          </person-group>
          <article-title>Pharmacological agents with effects on voice</article-title>
          <source>Am J Otolaryngol</source>
          <year>1995</year>
          <volume>16</volume>
          <issue>1</issue>
          <fpage>12</fpage>
          <lpage>8</lpage>
          <pub-id pub-id-type="doi">10.1016/0196-0709(95)90003-9</pub-id>
          <pub-id pub-id-type="medline">7717466</pub-id>
          <pub-id pub-id-type="pii">0196-0709(95)90003-9</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref9">
        <label>9</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Hoehn-Saric</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>McLeod</surname>
              <given-names>DR</given-names>
            </name>
          </person-group>
          <article-title>The peripheral sympathetic nervous system. Its role in normal and pathologic anxiety</article-title>
          <source>Psychiatr Clin North Am</source>
          <year>1988</year>
          <month>06</month>
          <volume>11</volume>
          <issue>2</issue>
          <fpage>375</fpage>
          <lpage>86</lpage>
          <pub-id pub-id-type="medline">3047706</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref10">
        <label>10</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Pennebaker</surname>
              <given-names>JW</given-names>
            </name>
            <name name-style="western">
              <surname>Mehl</surname>
              <given-names>MR</given-names>
            </name>
            <name name-style="western">
              <surname>Niederhoffer</surname>
              <given-names>KG</given-names>
            </name>
          </person-group>
          <article-title>Psychological aspects of natural language. Use: our words, our selves</article-title>
          <source>Annu Rev Psychol</source>
          <year>2003</year>
          <volume>54</volume>
          <fpage>547</fpage>
          <lpage>77</lpage>
          <pub-id pub-id-type="doi">10.1146/annurev.psych.54.101601.145041</pub-id>
          <pub-id pub-id-type="medline">12185209</pub-id>
          <pub-id pub-id-type="pii">101601.145041</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref11">
        <label>11</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Teferra</surname>
              <given-names>BG</given-names>
            </name>
            <name name-style="western">
              <surname>Borwein</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>DeSouza</surname>
              <given-names>DD</given-names>
            </name>
            <name name-style="western">
              <surname>Simpson</surname>
              <given-names>W</given-names>
            </name>
            <name name-style="western">
              <surname>Rheault</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Rose</surname>
              <given-names>J</given-names>
            </name>
          </person-group>
          <article-title>Acoustic and linguistic features of impromptu speech and their association with anxiety: validation study</article-title>
          <source>JMIR Ment Health</source>
          <year>2022</year>
          <month>07</month>
          <day>08</day>
          <volume>9</volume>
          <issue>7</issue>
          <fpage>e36828</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://mental.jmir.org/2022/7/e36828/"/>
          </comment>
          <pub-id pub-id-type="doi">10.2196/36828</pub-id>
          <pub-id pub-id-type="medline">35802401</pub-id>
          <pub-id pub-id-type="pii">v9i7e36828</pub-id>
          <pub-id pub-id-type="pmcid">PMC9308078</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref12">
        <label>12</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Spitzer</surname>
              <given-names>RL</given-names>
            </name>
            <name name-style="western">
              <surname>Kroenke</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Williams</surname>
              <given-names>JB</given-names>
            </name>
            <name name-style="western">
              <surname>Löwe</surname>
              <given-names>B</given-names>
            </name>
          </person-group>
          <article-title>A brief measure for assessing generalized anxiety disorder: the GAD-7</article-title>
          <source>Arch Intern Med</source>
          <year>2006</year>
          <month>05</month>
          <day>22</day>
          <volume>166</volume>
          <issue>10</issue>
          <fpage>1092</fpage>
          <lpage>7</lpage>
          <pub-id pub-id-type="doi">10.1001/archinte.166.10.1092</pub-id>
          <pub-id pub-id-type="medline">16717171</pub-id>
          <pub-id pub-id-type="pii">166/10/1092</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref13">
        <label>13</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>McGinnis</surname>
              <given-names>EW</given-names>
            </name>
            <name name-style="western">
              <surname>Anderau</surname>
              <given-names>SP</given-names>
            </name>
            <name name-style="western">
              <surname>Hruschak</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Gurchiek</surname>
              <given-names>RD</given-names>
            </name>
            <name name-style="western">
              <surname>Lopez-Duran</surname>
              <given-names>NL</given-names>
            </name>
            <name name-style="western">
              <surname>Fitzgerald</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Rosenblum</surname>
              <given-names>KL</given-names>
            </name>
            <name name-style="western">
              <surname>Muzik</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>McGinnis</surname>
              <given-names>RS</given-names>
            </name>
          </person-group>
          <article-title>Giving voice to vulnerable children: machine learning analysis of speech detects anxiety and depression in early childhood</article-title>
          <source>IEEE J Biomed Health Inform</source>
          <year>2019</year>
          <month>11</month>
          <volume>23</volume>
          <issue>6</issue>
          <fpage>2294</fpage>
          <lpage>301</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://europepmc.org/abstract/MED/31034426"/>
          </comment>
          <pub-id pub-id-type="doi">10.1109/JBHI.2019.2913590</pub-id>
          <pub-id pub-id-type="medline">31034426</pub-id>
          <pub-id pub-id-type="pmcid">PMC7484854</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref14">
        <label>14</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Buske-Kirschbaum</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Jobst</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Wustmans</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Kirschbaum</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Rauh</surname>
              <given-names>W</given-names>
            </name>
            <name name-style="western">
              <surname>Hellhammer</surname>
              <given-names>D</given-names>
            </name>
          </person-group>
          <article-title>Attenuated free cortisol response to psychosocial stress in children with atopic dermatitis</article-title>
          <source>Psychosom Med</source>
          <year>1997</year>
          <volume>59</volume>
          <issue>4</issue>
          <fpage>419</fpage>
          <lpage>26</lpage>
          <pub-id pub-id-type="doi">10.1097/00006842-199707000-00012</pub-id>
          <pub-id pub-id-type="medline">9251162</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref15">
        <label>15</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Ali</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Tanweer</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Khalid</surname>
              <given-names>SS</given-names>
            </name>
            <name name-style="western">
              <surname>Rao</surname>
              <given-names>N</given-names>
            </name>
          </person-group>
          <article-title>Mel frequency cepstral coefficient: a review</article-title>
          <source>Proceedings of the 2nd International Conference on ICT for Digital, Smart, and Sustainable Development</source>
          <year>2020</year>
          <conf-name>ICIDSSD '20</conf-name>
          <conf-date>February 27-28, 2020</conf-date>
          <conf-loc>New Delhi, India</conf-loc>
          <pub-id pub-id-type="doi">10.4108/eai.27-2-2020.2303173</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref16">
        <label>16</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Cortes</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Vapnik</surname>
              <given-names>V</given-names>
            </name>
          </person-group>
          <article-title>Support-vector networks</article-title>
          <source>Mach Learn</source>
          <year>1995</year>
          <month>9</month>
          <volume>20</volume>
          <issue>3</issue>
          <fpage>273</fpage>
          <lpage>97</lpage>
          <pub-id pub-id-type="doi">10.1007/BF00994018</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref17">
        <label>17</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Weeks</surname>
              <given-names>JW</given-names>
            </name>
            <name name-style="western">
              <surname>Lee</surname>
              <given-names>CY</given-names>
            </name>
            <name name-style="western">
              <surname>Reilly</surname>
              <given-names>AR</given-names>
            </name>
            <name name-style="western">
              <surname>Howell</surname>
              <given-names>AN</given-names>
            </name>
            <name name-style="western">
              <surname>France</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Kowalsky</surname>
              <given-names>JM</given-names>
            </name>
            <name name-style="western">
              <surname>Bush</surname>
              <given-names>A</given-names>
            </name>
          </person-group>
          <article-title>"The Sound of Fear": assessing vocal fundamental frequency as a physiological indicator of social anxiety disorder</article-title>
          <source>J Anxiety Disord</source>
          <year>2012</year>
          <month>12</month>
          <volume>26</volume>
          <issue>8</issue>
          <fpage>811</fpage>
          <lpage>22</lpage>
          <pub-id pub-id-type="doi">10.1016/j.janxdis.2012.07.005</pub-id>
          <pub-id pub-id-type="medline">23070030</pub-id>
          <pub-id pub-id-type="pii">S0887-6185(12)00093-X</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref18">
        <label>18</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Julian</surname>
              <given-names>LJ</given-names>
            </name>
          </person-group>
          <article-title>Measures of anxiety: State-Trait Anxiety Inventory (STAI), Beck Anxiety Inventory (BAI), and Hospital Anxiety and Depression Scale-Anxiety (HADS-A)</article-title>
          <source>Arthritis Care Res (Hoboken)</source>
          <year>2011</year>
          <month>11</month>
          <volume>63 Suppl 11</volume>
          <fpage>S467</fpage>
          <lpage>72</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.1002/acr.20561"/>
          </comment>
          <pub-id pub-id-type="doi">10.1002/acr.20561</pub-id>
          <pub-id pub-id-type="medline">22588767</pub-id>
          <pub-id pub-id-type="pmcid">PMC3879951</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref19">
        <label>19</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Weeks</surname>
              <given-names>JW</given-names>
            </name>
            <name name-style="western">
              <surname>Srivastav</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Howell</surname>
              <given-names>AN</given-names>
            </name>
            <name name-style="western">
              <surname>Menatti</surname>
              <given-names>AR</given-names>
            </name>
          </person-group>
          <article-title>“Speaking more than words”: classifying men with social anxiety disorder via vocal acoustic analyses of diagnostic interviews</article-title>
          <source>J Psychopathol Behav Assess</source>
          <year>2015</year>
          <month>5</month>
          <day>30</day>
          <volume>38</volume>
          <issue>1</issue>
          <fpage>30</fpage>
          <lpage>41</lpage>
          <pub-id pub-id-type="doi">10.1007/s10862-015-9495-9</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref20">
        <label>20</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Salekin</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Eberle</surname>
              <given-names>JW</given-names>
            </name>
            <name name-style="western">
              <surname>Glenn</surname>
              <given-names>JJ</given-names>
            </name>
            <name name-style="western">
              <surname>Teachman</surname>
              <given-names>BA</given-names>
            </name>
            <name name-style="western">
              <surname>Stankovic</surname>
              <given-names>JA</given-names>
            </name>
          </person-group>
          <article-title>A weakly supervised learning framework for detecting social anxiety and depression</article-title>
          <source>Proc ACM Interact Mob Wearable Ubiquitous Technol</source>
          <year>2018</year>
          <month>06</month>
          <volume>2</volume>
          <issue>2</issue>
          <fpage>81</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://europepmc.org/abstract/MED/31187083"/>
          </comment>
          <pub-id pub-id-type="doi">10.1145/3214284</pub-id>
          <pub-id pub-id-type="medline">31187083</pub-id>
          <pub-id pub-id-type="pmcid">PMC6559734</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref21">
        <label>21</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Baird</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Cummins</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Schnieder</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Krajewski</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Schuller</surname>
              <given-names>BW</given-names>
            </name>
          </person-group>
          <article-title>An evaluation of the effect of anxiety on speech — computational prediction of anxiety from sustained vowels</article-title>
          <source>Proceedings of 2020 Interspeech</source>
          <year>2020</year>
          <conf-name>InterSpeech '20</conf-name>
          <conf-date>October 25-29, 2020</conf-date>
          <conf-loc>Shanghai, China</conf-loc>
          <fpage>4951</fpage>
          <lpage>5</lpage>
          <pub-id pub-id-type="doi">10.21437/interspeech.2020-1801</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref22">
        <label>22</label>
        <nlm-citation citation-type="book">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Beck</surname>
              <given-names>AT</given-names>
            </name>
            <name name-style="western">
              <surname>Epstein</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Brown</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>Steer</surname>
              <given-names>R</given-names>
            </name>
          </person-group>
          <source>Beck Anxiety Inventory</source>
          <year>1988</year>
          <publisher-loc>Washington, DC, USA</publisher-loc>
          <publisher-name>American Psychological Association</publisher-name>
        </nlm-citation>
      </ref>
      <ref id="ref23">
        <label>23</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Rook</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Mazza</surname>
              <given-names>MC</given-names>
            </name>
            <name name-style="western">
              <surname>Lefter</surname>
              <given-names>I</given-names>
            </name>
            <name name-style="western">
              <surname>Brazier</surname>
              <given-names>F</given-names>
            </name>
          </person-group>
          <article-title>Toward linguistic recognition of generalized anxiety disorder</article-title>
          <source>Front Digit Health</source>
          <year>2022</year>
          <month>4</month>
          <day>15</day>
          <volume>4</volume>
          <fpage>779039</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://europepmc.org/abstract/MED/35493530"/>
          </comment>
          <pub-id pub-id-type="doi">10.3389/fdgth.2022.779039</pub-id>
          <pub-id pub-id-type="medline">35493530</pub-id>
          <pub-id pub-id-type="pmcid">PMC9051024</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref24">
        <label>24</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Carver</surname>
              <given-names>CS</given-names>
            </name>
            <name name-style="western">
              <surname>White</surname>
              <given-names>TL</given-names>
            </name>
          </person-group>
          <article-title>Behavioral inhibition, behavioral activation, and affective responses to impending reward and punishment: the BIS/BAS scales</article-title>
          <source>J Pers Soc Psychol</source>
          <year>1994</year>
          <month>08</month>
          <volume>67</volume>
          <issue>2</issue>
          <fpage>319</fpage>
          <lpage>33</lpage>
          <pub-id pub-id-type="doi">10.1037/0022-3514.67.2.319</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref25">
        <label>25</label>
        <nlm-citation citation-type="web">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Pennebaker</surname>
              <given-names>JW</given-names>
            </name>
            <name name-style="western">
              <surname>Boyd</surname>
              <given-names>RL</given-names>
            </name>
            <name name-style="western">
              <surname>Jordan</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Blackburn</surname>
              <given-names>K</given-names>
            </name>
          </person-group>
          <article-title>The development and psychometric properties of LIWC2015</article-title>
          <source>The University of Texas at Austin</source>
          <year>2015</year>
          <access-date>2022-01-24</access-date>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://repositories.lib.utexas.edu/bitstream/handle/2152/31333/LIWC2015_LanguageManual.pdf?sequenc">https://repositories.lib.utexas.edu/bitstream/handle/2152/31333/LIWC2015_LanguageManual.pdf?sequenc</ext-link>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref26">
        <label>26</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Di Matteo</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Fotinos</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Lokuge</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Yu</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Sternat</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Katzman</surname>
              <given-names>MA</given-names>
            </name>
            <name name-style="western">
              <surname>Rose</surname>
              <given-names>J</given-names>
            </name>
          </person-group>
          <article-title>The relationship between smartphone-recorded environmental audio and symptomatology of anxiety and depression: exploratory study</article-title>
          <source>JMIR Form Res</source>
          <year>2020</year>
          <month>08</month>
          <day>13</day>
          <volume>4</volume>
          <issue>8</issue>
          <fpage>e18751</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://formative.jmir.org/2020/8/e18751/"/>
          </comment>
          <pub-id pub-id-type="doi">10.2196/18751</pub-id>
          <pub-id pub-id-type="medline">32788153</pub-id>
          <pub-id pub-id-type="pii">v4i8e18751</pub-id>
          <pub-id pub-id-type="pmcid">PMC7453326</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref27">
        <label>27</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Di Matteo</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>W</given-names>
            </name>
            <name name-style="western">
              <surname>Fotinos</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Lokuge</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Yu</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Sternat</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Katzman</surname>
              <given-names>MA</given-names>
            </name>
            <name name-style="western">
              <surname>Rose</surname>
              <given-names>J</given-names>
            </name>
          </person-group>
          <article-title>Smartphone-detected ambient speech and self-reported measures of anxiety and depression: exploratory observational study</article-title>
          <source>JMIR Form Res</source>
          <year>2021</year>
          <month>01</month>
          <day>29</day>
          <volume>5</volume>
          <issue>1</issue>
          <fpage>e22723</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://formative.jmir.org/2021/1/e22723/"/>
          </comment>
          <pub-id pub-id-type="doi">10.2196/22723</pub-id>
          <pub-id pub-id-type="medline">33512325</pub-id>
          <pub-id pub-id-type="pii">v5i1e22723</pub-id>
          <pub-id pub-id-type="pmcid">PMC7880807</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref28">
        <label>28</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Di Matteo</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Fotinos</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Lokuge</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Mason</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>Sternat</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Katzman</surname>
              <given-names>MA</given-names>
            </name>
            <name name-style="western">
              <surname>Rose</surname>
              <given-names>J</given-names>
            </name>
          </person-group>
          <article-title>Automated screening for social anxiety, generalized anxiety, and depression from objective smartphone-collected data: cross-sectional study</article-title>
          <source>J Med Internet Res</source>
          <year>2021</year>
          <month>08</month>
          <day>13</day>
          <volume>23</volume>
          <issue>8</issue>
          <fpage>e28918</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.jmir.org/2021/8/e28918/"/>
          </comment>
          <pub-id pub-id-type="doi">10.2196/28918</pub-id>
          <pub-id pub-id-type="medline">34397386</pub-id>
          <pub-id pub-id-type="pii">v23i8e28918</pub-id>
          <pub-id pub-id-type="pmcid">PMC8398720</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref29">
        <label>29</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Palan</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Schitter</surname>
              <given-names>C</given-names>
            </name>
          </person-group>
          <article-title>Prolific.ac—a subject pool for online experiments</article-title>
          <source>J Behav Exp Finance</source>
          <year>2018</year>
          <month>03</month>
          <volume>17</volume>
          <fpage>22</fpage>
          <lpage>7</lpage>
          <pub-id pub-id-type="doi">10.1016/j.jbef.2017.12.004</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref30">
        <label>30</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Kirschbaum</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Pirke</surname>
              <given-names>KM</given-names>
            </name>
            <name name-style="western">
              <surname>Hellhammer</surname>
              <given-names>DH</given-names>
            </name>
          </person-group>
          <article-title>The 'Trier Social Stress Test'--a tool for investigating psychobiological stress responses in a laboratory setting</article-title>
          <source>Neuropsychobiology</source>
          <year>1993</year>
          <volume>28</volume>
          <issue>1-2</issue>
          <fpage>76</fpage>
          <lpage>81</lpage>
          <pub-id pub-id-type="doi">10.1159/000119004</pub-id>
          <pub-id pub-id-type="medline">8255414</pub-id>
          <pub-id pub-id-type="pii">119004</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref31">
        <label>31</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Gerra</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>Zaimovic</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Zambelli</surname>
              <given-names>U</given-names>
            </name>
            <name name-style="western">
              <surname>Timpano</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Reali</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Bernasconi</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Brambilla</surname>
              <given-names>F</given-names>
            </name>
          </person-group>
          <article-title>Neuroendocrine responses to psychological stress in adolescents with anxiety disorder</article-title>
          <source>Neuropsychobiology</source>
          <year>2000</year>
          <volume>42</volume>
          <issue>2</issue>
          <fpage>82</fpage>
          <lpage>92</lpage>
          <pub-id pub-id-type="doi">10.1159/000026677</pub-id>
          <pub-id pub-id-type="medline">10940763</pub-id>
          <pub-id pub-id-type="pii">26677</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref32">
        <label>32</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Jezova</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Makatsori</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Duncko</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Moncek</surname>
              <given-names>F</given-names>
            </name>
            <name name-style="western">
              <surname>Jakubek</surname>
              <given-names>M</given-names>
            </name>
          </person-group>
          <article-title>High trait anxiety in healthy subjects is associated with low neuroendocrine activity during psychosocial stress</article-title>
          <source>Prog Neuropsychopharmacol Biol Psychiatry</source>
          <year>2004</year>
          <month>12</month>
          <volume>28</volume>
          <issue>8</issue>
          <fpage>1331</fpage>
          <lpage>6</lpage>
          <pub-id pub-id-type="doi">10.1016/j.pnpbp.2004.08.005</pub-id>
          <pub-id pub-id-type="medline">15588760</pub-id>
          <pub-id pub-id-type="pii">S0278-5846(04)00183-6</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref33">
        <label>33</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Endler</surname>
              <given-names>NS</given-names>
            </name>
            <name name-style="western">
              <surname>Kocovski</surname>
              <given-names>NL</given-names>
            </name>
          </person-group>
          <article-title>State and trait anxiety revisited</article-title>
          <source>J Anxiety Disord</source>
          <year>2001</year>
          <volume>15</volume>
          <issue>3</issue>
          <fpage>231</fpage>
          <lpage>45</lpage>
          <pub-id pub-id-type="doi">10.1016/s0887-6185(01)00060-3</pub-id>
          <pub-id pub-id-type="medline">11442141</pub-id>
          <pub-id pub-id-type="pii">S0887-6185(01)00060-3</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref34">
        <label>34</label>
        <nlm-citation citation-type="book">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>First</surname>
              <given-names>MB</given-names>
            </name>
            <name name-style="western">
              <surname>Spitzer</surname>
              <given-names>RL</given-names>
            </name>
            <name name-style="western">
              <surname>Gibbon</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Williams</surname>
              <given-names>JB</given-names>
            </name>
          </person-group>
          <source>Structured Clinical Interview for DSM-IV Axis I Disorders (SCID-I), Clinician Version</source>
          <year>1997</year>
          <publisher-loc>Washington, DC, USA</publisher-loc>
          <publisher-name>American Psychiatric Press</publisher-name>
        </nlm-citation>
      </ref>
      <ref id="ref35">
        <label>35</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Cameron</surname>
              <given-names>OG</given-names>
            </name>
            <name name-style="western">
              <surname>Hill</surname>
              <given-names>EM</given-names>
            </name>
          </person-group>
          <article-title>Women and anxiety</article-title>
          <source>Psychiatr Clin North Am</source>
          <year>1989</year>
          <month>03</month>
          <volume>12</volume>
          <issue>1</issue>
          <fpage>175</fpage>
          <lpage>86</lpage>
          <pub-id pub-id-type="doi">10.1016/s0193-953x(18)30459-3</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref36">
        <label>36</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Krasucki</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Howard</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Mann</surname>
              <given-names>A</given-names>
            </name>
          </person-group>
          <article-title>The relationship between anxiety disorders and age</article-title>
          <source>Int J Geriatr Psychiatry</source>
          <year>1998</year>
          <month>02</month>
          <volume>13</volume>
          <issue>2</issue>
          <fpage>79</fpage>
          <lpage>99</lpage>
          <pub-id pub-id-type="doi">10.1002/(sici)1099-1166(199802)13:2&#60;79::aid-gps739&#62;3.0.co;2-g</pub-id>
          <pub-id pub-id-type="medline">9526178</pub-id>
          <pub-id pub-id-type="pii">10.1002/(SICI)1099-1166(199802)13:2&#60;79::AID-GPS739&#62;3.0.CO;2-G</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref37">
        <label>37</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Dijkstra-Kersten</surname>
              <given-names>SM</given-names>
            </name>
            <name name-style="western">
              <surname>Biesheuvel-Leliefeld</surname>
              <given-names>KE</given-names>
            </name>
            <name name-style="western">
              <surname>van der Wouden</surname>
              <given-names>JC</given-names>
            </name>
            <name name-style="western">
              <surname>Penninx</surname>
              <given-names>BW</given-names>
            </name>
            <name name-style="western">
              <surname>van Marwijk</surname>
              <given-names>HW</given-names>
            </name>
          </person-group>
          <article-title>Associations of financial strain and income with depressive and anxiety disorders</article-title>
          <source>J Epidemiol Community Health</source>
          <year>2015</year>
          <month>07</month>
          <volume>69</volume>
          <issue>7</issue>
          <fpage>660</fpage>
          <lpage>5</lpage>
          <pub-id pub-id-type="doi">10.1136/jech-2014-205088</pub-id>
          <pub-id pub-id-type="medline">25636322</pub-id>
          <pub-id pub-id-type="pii">jech-2014-205088</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref38">
        <label>38</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Bischl</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Mersmann</surname>
              <given-names>O</given-names>
            </name>
            <name name-style="western">
              <surname>Trautmann</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Weihs</surname>
              <given-names>C</given-names>
            </name>
          </person-group>
          <article-title>Resampling methods for meta-model validation with recommendations for evolutionary computation</article-title>
          <source>Evol Comput</source>
          <year>2012</year>
          <volume>20</volume>
          <issue>2</issue>
          <fpage>249</fpage>
          <lpage>75</lpage>
          <pub-id pub-id-type="doi">10.1162/EVCO_a_00069</pub-id>
          <pub-id pub-id-type="medline">22339368</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref39">
        <label>39</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Belkin</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Hsu</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Ma</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Mandal</surname>
              <given-names>S</given-names>
            </name>
          </person-group>
          <article-title>Reconciling modern machine-learning practice and the classical bias-variance trade-off</article-title>
          <source>Proc Natl Acad Sci U S A</source>
          <year>2019</year>
          <month>08</month>
          <day>06</day>
          <volume>116</volume>
          <issue>32</issue>
          <fpage>15849</fpage>
          <lpage>54</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://europepmc.org/abstract/MED/31341078"/>
          </comment>
          <pub-id pub-id-type="doi">10.1073/pnas.1903070116</pub-id>
          <pub-id pub-id-type="medline">31341078</pub-id>
          <pub-id pub-id-type="pii">1903070116</pub-id>
          <pub-id pub-id-type="pmcid">PMC6689936</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref40">
        <label>40</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Bouckaert</surname>
              <given-names>RR</given-names>
            </name>
            <name name-style="western">
              <surname>Frank</surname>
              <given-names>E</given-names>
            </name>
          </person-group>
          <article-title>Evaluating the replicability of significance tests for comparing learning algorithms</article-title>
          <source>Proceedings of the 8th Pacific-Asia Conference on Advances in Knowledge Discovery and Data Mining</source>
          <year>2004</year>
          <conf-name>PAKDD '04</conf-name>
          <conf-date>May 26-28, 2004</conf-date>
          <conf-loc>Sydney, Australia</conf-loc>
          <fpage>3</fpage>
          <lpage>12</lpage>
          <pub-id pub-id-type="doi">10.1007/978-3-540-24775-3_3</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref41">
        <label>41</label>
        <nlm-citation citation-type="web">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Di Matteo</surname>
              <given-names>D</given-names>
            </name>
          </person-group>
          <article-title>Inference of anxiety and depression from smartphone-collected data</article-title>
          <source>University of Toronto</source>
          <year>2021</year>
          <access-date>2022-01-24</access-date>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://tspace.library.utoronto.ca/bitstream/1807/108894/1/Di_Matteo_Daniel_202111_PhD_thesis.pdf">https://tspace.library.utoronto.ca/bitstream/1807/108894/1/Di_Matteo_Daniel_202111_PhD_thesis.pdf</ext-link>
          </comment>
        </nlm-citation>
      </ref>
    </ref-list>
  </back>
</article>
