<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.0 20040830//EN" "http://dtd.nlm.nih.gov/publishing/2.0/journalpublishing.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink" article-type="research-article" dtd-version="2.0">
  <front>
    <journal-meta>
      <journal-id journal-id-type="publisher-id">JFR</journal-id>
      <journal-id journal-id-type="nlm-ta">JMIR Form Res</journal-id>
      <journal-title>JMIR Formative Research</journal-title>
      <issn pub-type="epub">2561-326X</issn>
      <publisher>
        <publisher-name>JMIR Publications</publisher-name>
        <publisher-loc>Toronto, Canada</publisher-loc>
      </publisher>
    </journal-meta>
    <article-meta>
      <article-id pub-id-type="publisher-id">v6i8e38092</article-id>
      <article-id pub-id-type="pmid">35969431</article-id>
      <article-id pub-id-type="doi">10.2196/38092</article-id>
      <article-categories>
        <subj-group subj-group-type="heading">
          <subject>Original Paper</subject>
        </subj-group>
        <subj-group subj-group-type="article-type">
          <subject>Original Paper</subject>
        </subj-group>
      </article-categories>
      <title-group>
        <article-title>Design and Formative Evaluation of a Virtual Voice-Based Coach for Problem-solving Treatment: Observational Study</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="editor">
          <name>
            <surname>Mavragani</surname>
            <given-names>Amaryllis</given-names>
          </name>
        </contrib>
      </contrib-group>
      <contrib-group>
        <contrib contrib-type="reviewer">
          <name>
            <surname>Bhattacharjee</surname>
            <given-names>Ananya</given-names>
          </name>
        </contrib>
        <contrib contrib-type="reviewer">
          <name>
            <surname>Pritchard</surname>
            <given-names>Michael</given-names>
          </name>
        </contrib>
      </contrib-group>
      <contrib-group>
        <contrib id="contrib1" contrib-type="author">
          <name name-style="western">
            <surname>Kannampallil</surname>
            <given-names>Thomas</given-names>
          </name>
          <degrees>PhD</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0003-4119-4836</ext-link>
        </contrib>
        <contrib id="contrib2" contrib-type="author" equal-contrib="yes">
          <name name-style="western">
            <surname>Ronneberg</surname>
            <given-names>Corina R</given-names>
          </name>
          <degrees>MS</degrees>
          <xref rid="aff2" ref-type="aff">2</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0001-8105-1886</ext-link>
        </contrib>
        <contrib id="contrib3" contrib-type="author" equal-contrib="yes">
          <name name-style="western">
            <surname>Wittels</surname>
            <given-names>Nancy E</given-names>
          </name>
          <degrees>MS</degrees>
          <xref rid="aff2" ref-type="aff">2</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0002-0778-2573</ext-link>
        </contrib>
        <contrib id="contrib4" contrib-type="author">
          <name name-style="western">
            <surname>Kumar</surname>
            <given-names>Vikas</given-names>
          </name>
          <degrees>MS</degrees>
          <xref rid="aff2" ref-type="aff">2</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0002-5490-7252</ext-link>
        </contrib>
        <contrib id="contrib5" contrib-type="author">
          <name name-style="western">
            <surname>Lv</surname>
            <given-names>Nan</given-names>
          </name>
          <degrees>PhD</degrees>
          <xref rid="aff2" ref-type="aff">2</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0003-2063-3377</ext-link>
        </contrib>
        <contrib id="contrib6" contrib-type="author">
          <name name-style="western">
            <surname>Smyth</surname>
            <given-names>Joshua M</given-names>
          </name>
          <degrees>PhD</degrees>
          <xref rid="aff3" ref-type="aff">3</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0002-0904-5390</ext-link>
        </contrib>
        <contrib id="contrib7" contrib-type="author">
          <name name-style="western">
            <surname>Gerber</surname>
            <given-names>Ben S</given-names>
          </name>
          <degrees>MPH, MD</degrees>
          <xref rid="aff4" ref-type="aff">4</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0003-4367-6396</ext-link>
        </contrib>
        <contrib id="contrib8" contrib-type="author">
          <name name-style="western">
            <surname>Kringle</surname>
            <given-names>Emily A</given-names>
          </name>
          <degrees>PhD</degrees>
          <xref rid="aff2" ref-type="aff">2</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0003-4958-9568</ext-link>
        </contrib>
        <contrib id="contrib9" contrib-type="author">
          <name name-style="western">
            <surname>Johnson</surname>
            <given-names>Jillian A</given-names>
          </name>
          <degrees>PhD</degrees>
          <xref rid="aff3" ref-type="aff">3</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0003-1379-2150</ext-link>
        </contrib>
        <contrib id="contrib10" contrib-type="author">
          <name name-style="western">
            <surname>Yu</surname>
            <given-names>Philip</given-names>
          </name>
          <degrees>PhD</degrees>
          <xref rid="aff2" ref-type="aff">2</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0002-3491-5968</ext-link>
        </contrib>
        <contrib id="contrib11" contrib-type="author">
          <name name-style="western">
            <surname>Steinman</surname>
            <given-names>Lesley E</given-names>
          </name>
          <degrees>MPH, MSW</degrees>
          <xref rid="aff5" ref-type="aff">5</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0002-2098-5105</ext-link>
        </contrib>
        <contrib id="contrib12" contrib-type="author" equal-contrib="yes">
          <name name-style="western">
            <surname>Ajilore</surname>
            <given-names>Olu A</given-names>
          </name>
          <degrees>MD, PhD</degrees>
          <xref rid="aff2" ref-type="aff">2</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0003-0737-0437</ext-link>
        </contrib>
        <contrib id="contrib13" contrib-type="author" corresp="yes" equal-contrib="yes">
          <name name-style="western">
            <surname>Ma</surname>
            <given-names>Jun</given-names>
          </name>
          <degrees>MD, PhD</degrees>
          <xref rid="aff2" ref-type="aff">2</xref>
          <address>
            <institution>University of Illinois at Chicago</institution>
            <addr-line>1747 W. Roosevelt Rd</addr-line>
            <addr-line>Room 466 (MC 275)</addr-line>
            <addr-line>Chicago, IL, 60608</addr-line>
            <country>United States</country>
            <phone>1 (312) 413 9830</phone>
            <email>maj2015@uic.edu</email>
          </address>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0001-7996-6454</ext-link>
        </contrib>
      </contrib-group>
      <aff id="aff1">
        <label>1</label>
        <institution>Washington University School of Medicine</institution>
        <addr-line>St. Louis, MO</addr-line>
        <country>United States</country>
      </aff>
      <aff id="aff2">
        <label>2</label>
        <institution>University of Illinois at Chicago</institution>
        <addr-line>Chicago, IL</addr-line>
        <country>United States</country>
      </aff>
      <aff id="aff3">
        <label>3</label>
        <institution>The Pennsylvania State University</institution>
        <addr-line>University Park, PA</addr-line>
        <country>United States</country>
      </aff>
      <aff id="aff4">
        <label>4</label>
        <institution>University of Massachusetts Medical School</institution>
        <addr-line>Worcester, MA</addr-line>
        <country>United States</country>
      </aff>
      <aff id="aff5">
        <label>5</label>
        <institution>University of Washington</institution>
        <addr-line>Seattle, WA</addr-line>
        <country>United States</country>
      </aff>
      <author-notes>
        <corresp>Corresponding Author: Jun Ma <email>maj2015@uic.edu</email></corresp>
      </author-notes>
      <pub-date pub-type="collection">
        <month>8</month>
        <year>2022</year>
      </pub-date>
      <pub-date pub-type="epub">
        <day>12</day>
        <month>8</month>
        <year>2022</year>
      </pub-date>
      <volume>6</volume>
      <issue>8</issue>
      <elocation-id>e38092</elocation-id>
      <history>
        <date date-type="received">
          <day>18</day>
          <month>3</month>
          <year>2022</year>
        </date>
        <date date-type="rev-request">
          <day>15</day>
          <month>6</month>
          <year>2022</year>
        </date>
        <date date-type="rev-recd">
          <day>1</day>
          <month>7</month>
          <year>2022</year>
        </date>
        <date date-type="accepted">
          <day>4</day>
          <month>7</month>
          <year>2022</year>
        </date>
      </history>
      <copyright-statement>©Thomas Kannampallil, Corina R Ronneberg, Nancy E Wittels, Vikas Kumar, Nan Lv, Joshua M Smyth, Ben S Gerber, Emily A Kringle, Jillian A Johnson, Philip Yu, Lesley E Steinman, Olu A Ajilore, Jun Ma. Originally published in JMIR Formative Research (https://formative.jmir.org), 12.08.2022.</copyright-statement>
      <copyright-year>2022</copyright-year>
      <license license-type="open-access" xlink:href="https://creativecommons.org/licenses/by/4.0/">
        <p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR Formative Research, is properly cited. The complete bibliographic information, a link to the original publication on https://formative.jmir.org, as well as this copyright and license information must be included.</p>
      </license>
      <self-uri xlink:href="https://formative.jmir.org/2022/8/e38092" xlink:type="simple"/>
      <abstract>
        <sec sec-type="background">
          <title>Background</title>
          <p>Artificial intelligence has provided new opportunities for human interactions with technology for the practice of medicine. Among the recent artificial intelligence innovations, personal voice assistants have been broadly adopted. This highlights their potential for health care–related applications such as behavioral counseling to promote healthy lifestyle habits and emotional well-being. However, the use of voice-based applications for behavioral therapy has not been previously evaluated.</p>
        </sec>
        <sec sec-type="objective">
          <title>Objective</title>
          <p>This study aimed to conduct a formative user evaluation of Lumen, a virtual voice-based coach developed as an Alexa skill that delivers evidence-based, problem-solving treatment for patients with mild to moderate depression and/or anxiety.</p>
        </sec>
        <sec sec-type="methods">
          <title>Methods</title>
          <p>A total of 26 participants completed 2 therapy sessions—an introductory (session 1) and a problem-solving (session 2)—with Lumen. Following each session with Lumen, participants completed user experience, task-related workload, and work alliance surveys. They also participated in semistructured interviews addressing the benefits, challenges and barriers to Lumen use, and design recommendations. We evaluated the differences in user experience, task load, and work alliance between sessions using 2-tailed paired <italic>t</italic> tests. Interview transcripts were coded using an inductive thematic analysis to characterize the participants’ perspectives regarding Lumen use.</p>
        </sec>
        <sec sec-type="results">
          <title>Results</title>
          <p>Participants found Lumen to provide high pragmatic usability and favorable user experience, with marginal task load during interactions for both Lumen sessions. However, participants experienced a higher temporal workload during the problem-solving session, suggesting a feeling of being rushed during their communicative interactions. On the basis of the qualitative analysis, the following themes were identified: Lumen’s on-demand accessibility and the delivery of a complex problem-solving treatment task with a simplistic structure for achieving therapy goals; themes related to Lumen improvements included streamlining and improved personalization of conversations, slower pacing of conversations, and providing additional context during therapy sessions.</p>
        </sec>
        <sec sec-type="conclusions">
          <title>Conclusions</title>
          <p>On the basis of an in-depth formative evaluation, we found that Lumen supported the ability to conduct cognitively plausible interactions for the delivery of behavioral therapy. Several design suggestions identified from the study including reducing temporal and cognitive load during conversational interactions, developing more natural conversations, and expanding privacy and security features were incorporated in the revised version of Lumen. Although further research is needed, the promising findings from this study highlight the potential for using Lumen to deliver personalized and accessible mental health care, filling a gap in traditional mental health services.</p>
        </sec>
      </abstract>
      <kwd-group>
        <kwd>voice assistants</kwd>
        <kwd>behavioral therapy</kwd>
        <kwd>problem-solving therapy</kwd>
        <kwd>mental health</kwd>
        <kwd>artificial intelligence</kwd>
        <kwd>user evaluation</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec sec-type="introduction">
      <title>Introduction</title>
      <p>Artificial intelligence (AI) has provided new opportunities for human interactions with technology for care delivery [<xref ref-type="bibr" rid="ref1">1</xref>]. These include remote monitoring, mobile health apps (eg, chatbots), and the use of a wide variety of sensors for remote monitoring and surveillance. Of the recent innovations, personal voice assistants that rely on AI-based platforms such as Amazon’s Alexa, Google Home, Cortana, and Siri have transformed how humans search for information, with recent reports suggesting that nearly 30% of search queries rely on voice-based input [<xref ref-type="bibr" rid="ref2">2</xref>]. Broad adoption of such platforms lends support for their potential utility in health care–related applications such as behavioral counseling to promote healthy lifestyle habits and emotional well-being [<xref ref-type="bibr" rid="ref3">3</xref>,<xref ref-type="bibr" rid="ref4">4</xref>]. However, current health care–related applications of voice assistants are generally rudimentary, and few of them have been developed for delivering evidence-based therapies or have been subjected to careful evaluation (eg, to inform development or for their effect on clinical or behavioral outcomes) [<xref ref-type="bibr" rid="ref5">5</xref>]. To this end, we developed and evaluated Lumen, an end-to-end voice-based virtual coach that was developed as a stand-alone Alexa application. Lumen delivers evidence-based problem-solving treatment (PST) for patients with mild to moderate symptoms of depression and anxiety.</p>
      <p>Lumen, by design, is different from the current spectrum of voice-based health applications that primarily support web-based information–seeking activities [<xref ref-type="bibr" rid="ref4">4</xref>]. Studies on such information-seeking activities performed on voice assistants have focused on the quality and content of voice assistant responses for several topics including health behavior and lifestyle [<xref ref-type="bibr" rid="ref6">6</xref>,<xref ref-type="bibr" rid="ref7">7</xref>], mental health, interpersonal violence, addiction help [<xref ref-type="bibr" rid="ref8">8</xref>,<xref ref-type="bibr" rid="ref9">9</xref>], patient and consumer safety risks [<xref ref-type="bibr" rid="ref10">10</xref>], vaccines [<xref ref-type="bibr" rid="ref11">11</xref>], postpartum depression [<xref ref-type="bibr" rid="ref12">12</xref>], medication names [<xref ref-type="bibr" rid="ref13">13</xref>], and sexual health [<xref ref-type="bibr" rid="ref14">14</xref>]. The findings across these studies consistently highlight the shortcomings associated with the quality of the information retrieved during these voice-based searches. For example, Bickmore et al [<xref ref-type="bibr" rid="ref10">10</xref>] found that Siri, Alexa, and Google Assistant platforms and their underlying algorithms were effective in completing only 43% of requests regarding situations that required medical expertise, and 29% of the responses could have resulted in some degree of patient harm [<xref ref-type="bibr" rid="ref10">10</xref>]. Other applications, mostly preliminary prototypes, have been developed for assessment and support. These applications have been used for delivering visual acuity tests [<xref ref-type="bibr" rid="ref15">15</xref>], support for coping with chronic disease [<xref ref-type="bibr" rid="ref16">16</xref>], and for nutritional planning [<xref ref-type="bibr" rid="ref17">17</xref>]. However, it is important to note that these applications have largely lacked outcome assessment or incorporation of behavioral therapy [<xref ref-type="bibr" rid="ref4">4</xref>]. Although text-based behavioral therapy applications (eg, chatbots) have shown promise in mitigating psychiatric disorders [<xref ref-type="bibr" rid="ref18">18</xref>,<xref ref-type="bibr" rid="ref19">19</xref>], several challenges exist including long-term adherence and engagement limited to younger age groups [<xref ref-type="bibr" rid="ref20">20</xref>]. Therefore, it is plausible that voice-based therapy delivery may mitigate some of these issues.</p>
      <p>In this paper, we describe the design and formative evaluation of Lumen, with the following research objectives: (1) to characterize the user experience, task-related workload associated with interactive communication, and participant alliance with delivered treatment and (2) to identify and describe user perspectives including the benefits, challenges, and barriers to Lumen use and recommendations for design improvements.</p>
    </sec>
    <sec sec-type="methods">
      <title>Methods</title>
      <p>In the following sections, we describe the design components of Lumen, its features, and the mixed methods study that was conducted.</p>
      <sec>
        <title>Lumen</title>
        <p>Lumen is a virtual voice-based coach that delivers an evidence-based, 8-session PST program for patients with mild to moderate depression and anxiety. The first 4 PST sessions were conducted weekly, followed by 4 biweekly sessions. Each PST session lasted approximately 45 minutes to 1 hour. Lumen was designed to align with the evidence-based PST program.</p>
        <p>Lumen’s design was based on two overarching principles: (1) providing cognitively plausible conversations, that is, aligning Lumen’s conversations with the cognitive processes of human communicative interactions [<xref ref-type="bibr" rid="ref5">5</xref>] and (2) alignment with the principles of evidence-based PST. This PST program was previously tested and delivered with a human coach [<xref ref-type="bibr" rid="ref21">21</xref>]; Lumen incorporates essential components of the treatment protocol for coaching and monitors progress using surveys and ecological momentary assessments. All Lumen design components are delivered in an integrated environment, coordinated through the voice-only platform and associated mobile tools (<xref rid="figure1" ref-type="fig">Figure 1</xref> provides an overview of the components of Lumen and their interactions).</p>
        <p>Developed on Amazon’s Alexa platform, Lumen’s architecture incorporates an intelligent conversation manager that manages the content, structure, and flow of interactive conversations between a patient and Lumen and a context manager that incorporates context awareness into the conversations. Using underlying AI capabilities of the Alexa platform, the conversation manager uses user verbal input to provide appropriate, synchronous responses, aligned with PST’s treatment guidelines. PST content and conversational structure were designed in consultation with master PST trainers and PST experts.</p>
        <p>The context manager provides contextual awareness to the interactions by incorporating user input from surveys and ecological momentary assessments (delivered asynchronously through mobile apps) and treatment progression and continuity (eg, review of patient problems and action plans from a previous session; Sections A and B in <xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref> provide additional details of the Lumen architecture and features).</p>
        <p>We followed an iterative user-centered design process, comprising brainstorming sessions with software engineers, interaction designers, psychiatrists, and researchers; prototype development on the Alexa platform; and several iterations of internal testing.</p>
        <fig id="figure1" position="float">
          <label>Figure 1</label>
          <caption>
            <p>User interaction with Lumen for problem-solving treatment (PST) sessions highlighting the various components. AWS: Amazon Web Services; EMA: ecological momentary assessment.</p>
          </caption>
          <graphic xlink:href="formative_v6i8e38092_fig1.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
      </sec>
      <sec>
        <title>Participants and Study Design</title>
        <p>Participants for this formative evaluation were recruited from the recently completed Engaging Self-Regulation Targets to Understand the Mechanisms of Behavior Change and Improve Mood and Weight Outcomes (ENGAGE-2) trial (ClinicalTrials.gov, National Clinical Trial#03841682), in which a PST-certified health coach delivered integrated collaborative care for depression and obesity to intervention participants, whereas those in the control group received usual care. A convenience sample (91/106, 85.8%) of ENGAGE-2 participants was contacted for assessing their interest in participating in a study with a virtual PST coach. Of these 91 participants, 26 (28%) expressed interest and consented to participate. Of the 26 participants, 17 (65%) had prior PST experience (ie, part of the ENGAGE-2 intervention group) and 9 (35%) did not have prior PST experience (ie, part of the ENGAGE-2 control group).</p>
        <p>This was an observational study, with each participant completing 2 Lumen sessions: an <italic>introductory</italic> first session (termed S1; n=26) and a <italic>problem-solving</italic> second session (termed S2; n=24, missing 2 of the 9 ENGAGE-2 control participants). The 2 sessions represented the overarching structure of the 8-session, evidence-based PST evaluated in a previous trial [<xref ref-type="bibr" rid="ref21">21</xref>]; S1 represented an initial overview session, and S2 represented a problem-solving session that was repeated in sessions 2 to 8 during the evidence-based PST.</p>
        <p>In S1, Lumen provides a program overview, provides a detailed introduction to the PST process and behavioral activation, and guides the participant to create a list of problems to address in subsequent sessions. In S2, Lumen guides the participant through the steps of problem-solving: identifying a problem to address, setting a goal, brainstorming possible solutions, evaluating the pros and cons of each solution, selecting a solution to implement, and developing an action plan to carry out before the next session. S2 concludes with behavioral activation coaching, where Lumen assists participants with selecting a social, physical, and pleasant activity to partake before the next session.</p>
        <p>The full Lumen PST program included 6 more problem-solving sessions that followed the same structure as S2; this was the rationale for testing only 1 problem-solving session during this formative evaluation. As such, the purpose of the 2-session approach was to conduct a representative evaluation of all Lumen sessions and to evaluate whether there were differences in participant experience and interactions between the sessions.</p>
      </sec>
      <sec>
        <title>Ethics Approval</title>
        <p>The study was approved by the institutional review board of the University of Illinois (IRB#2020-0918). All participants provided written consent.</p>
      </sec>
      <sec>
        <title>Procedure</title>
        <p>Consented participants were provided access to the Lumen S1 and S2 skills via the Alexa application and were given instructions on how to enable the skills on the Alexa app on their personal phones or mobile devices. All user interviews were conducted remotely by a trained research coordinator using the Zoom (Zoom Video Communications) videoconferencing platform. Participants were first provided with a brief overview of the study purpose, and their access to the Lumen skill (designed as a private skill, which was available by invitation only) was verified. During the session, a research coordinator went through a list of tips to effectively communicate with Lumen and answer any questions. After this, participants were instructed to turn off their video, and audio recording via Zoom was enabled from this point. Participants then opened the Alexa app and said “Open Lumen Coach” to begin their Lumen session. During their Lumen sessions, the trained note taker took notes of any deviations from the session script or any technical problems.</p>
        <p>After each Lumen session, the coordinator followed a semistructured interview script that included the following components. First, participants were asked to walk through their interaction experience with Lumen during their completed session, reflecting on what worked, what did not, and challenges they faced. Although the same procedure was followed for both Lumen sessions, interview questions varied slightly from S1 to S2 to inquire about session-specific content. Interview questions after S1 focused on participants’ impressions of Lumen, suggestions for improving Lumen, evaluating the usefulness of tips on how to communicate with Lumen, and impressions of the PST overview. Interview questions after S2 included questions about participants’ impressions of Lumen that were different from S1, delivery of PST by Lumen, and factors affecting their likelihood of Lumen use in the future. S1 and S2 were conducted several days apart, and participants had access to the specific sessions only a day or so before the session.</p>
        <p>After the interviews were completed, participants were emailed a link to 3 brief postinterview surveys related to user experience, workload, and the collaborative relationship between the participant and Lumen (User Experience Questionnaire Short Version [UEQ-S] [<xref ref-type="bibr" rid="ref22">22</xref>], NASA Task Load Index [TLX] [<xref ref-type="bibr" rid="ref23">23</xref>], and Working Alliance Inventory–Technology Version [WAI-Tech] [<xref ref-type="bibr" rid="ref24">24</xref>]).</p>
        <p>Audio recordings of the interviews (26 for S1 and 24 for S2) were transcribed using the Trint audio transcription software for subsequent analysis. All (26/26, 100%) postinterview surveys were completed after S1, and 95% (23/24) postinterview surveys were completed after S2.</p>
      </sec>
      <sec>
        <title>Data Analysis</title>
        <p>Data analysis included coding of interview transcripts using thematic analyses and descriptive summaries of user experience, task load, and WAI-Tech surveys.</p>
        <sec>
          <title>Coding of Transcripts</title>
          <p>All interview transcripts were coded using an inductive thematic analysis to characterize the participants’ perspectives regarding their interaction with Lumen [<xref ref-type="bibr" rid="ref25">25</xref>] (Section E in <xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref> provides the interview guide). This approach involved the following stages: first, 2 coauthors (CRR and EAK) read the interview transcripts to familiarize themselves with the content. Next, a set of “open codes” was created to characterize the content and context discussed in the interviews (ie, inductive coding) [<xref ref-type="bibr" rid="ref26">26</xref>]. These initial codes were compared across the transcripts to identify repeated and interrelated subthemes. Similar subthemes were grouped over multiple review sessions to develop a set of 6 overarching themes. All responses were coded; some responses were assigned multiple codes, in an order of relevance; however, only the primary assigned code was used for all analyses. Two coauthors (EAK and CRR) independently coded a set of 5 transcripts with a high degree of interrater agreement (Cohen κ ranged from 0.83 to 1 with mean 0.93, SD 0.07). Discrepancies were resolved through discussions with the first author (TK). Subsequently, all remaining transcripts were coded.</p>
        </sec>
        <sec>
          <title>Surveys</title>
          <p>From the UEQ-S survey, pragmatic quality and hedonic quality scale values were calculated by rescaling the survey responses to the range −3 to 3 and calculating item means within each scale using the UEQ-S Data Analysis Tool [<xref ref-type="bibr" rid="ref27">27</xref>]. Pragmatic quality refers to the task- or goal-related interaction qualities (eg, efficiency, perspicuity, and dependability) that a user aims to reach when using the product. Hedonic quality refers to the aspects related to pleasure or fun (eg, stimulation and novelty) while using the product. Values &#60;−0.8 represent a negative evaluation, between −0.8 and 0.8 represent a neutral evaluation, and &#62;0.8 represent a positive evaluation on each scale.</p>
          <p>The NASA TLX rating sheet was administered assuming similar weights for each of the 5 task load items (except for physical demand, which was not considered, as it was irrelevant to Lumen): mental demand, temporal demand (eg, being rushed), effort, frustration, and performance. Each item was then rescaled to the range 5 to 100 by multiplying the raw score by 5.</p>
          <p>From the WAI-Tech survey, three 12-item subscale (task, goal, and bond) scores and an overall score were calculated as item means within each subscale. The task subscale reflected how responsive Lumen was to the participant’s focus or need; the goal subscale reflected the extent to which goals were important, mutual, and capable of being accomplished; and the bond subscale reflected the degree of mutual liking and attachment [<xref ref-type="bibr" rid="ref24">24</xref>]. A higher overall score reflected a more positive rating of the working alliance.</p>
          <p>Given that the 2 sessions focused on 2 primary structural components of PST sessions—a session overview and a problem-solving session—we compared whether there were differences in the user experience, task load, or work alliance between these sessions. To this end, scores on each of the scales between S1 and S2 were compared using paired <italic>t</italic> tests. Analyses were conducted using SAS (version 9.4; SAS Institute Inc); statistical significance was defined by 2-sided <italic>P</italic>&#60;.05. Additional analyses comparing PST-experienced and PST-naive participants can be found in Section F in <xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref>.</p>
        </sec>
      </sec>
    </sec>
    <sec sec-type="results">
      <title>Results</title>
      <sec>
        <title>General Characteristics</title>
        <p>Among the 26 participants, 20 (77%) were female, 19 (73%) were racial or ethnic minorities (n=13, 50% Black; n=6, 23% Hispanic) with an average age of 43.9 (SD 11.9) years, 10 (38%) had a high school or some college education, and 14 (54%) had an annual family income of &#60;US $55,000 (<xref ref-type="table" rid="table1">Table 1</xref>). Participants with previous PST experience (17/26, 65%) and those without previous PST experience (9/26, 35%) did not differ in age, race, income, or educational status, although 65% (11/17) of the participants with previous PST experience and 100% (9/9) of the participants without PST experience were female (<italic>P</italic>=.04).</p>
        <table-wrap position="float" id="table1">
          <label>Table 1</label>
          <caption>
            <p>Baseline characteristics by prior problem-solving treatment (PST) experience.</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="30"/>
            <col width="250"/>
            <col width="220"/>
            <col width="210"/>
            <col width="210"/>
            <col width="80"/>
            <thead>
              <tr valign="top">
                <td colspan="2">Characteristic</td>
                <td>All Lumen formative evaluation participants (N=26)</td>
                <td>Participants with prior PST experience (n=17)</td>
                <td>Participants without prior PST experience (n=9)</td>
                <td><italic>P</italic> value</td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td colspan="2">Age (years), mean (SD)</td>
                <td>43.9 (11.9)</td>
                <td>42.6 (13.2)</td>
                <td>46.3 (9.2)</td>
                <td>.46</td>
              </tr>
              <tr valign="top">
                <td colspan="2">Female, n (%)</td>
                <td>20 (77)</td>
                <td>11 (65)</td>
                <td>9 (100)</td>
                <td>.04</td>
              </tr>
              <tr valign="top">
                <td colspan="6">
                  <bold>Race or ethnicity, n (%)</bold>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>Non-Hispanic White</td>
                <td>4 (15)</td>
                <td>3 (18)</td>
                <td>1 (11)</td>
                <td>.34</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>African American</td>
                <td>13 (50)</td>
                <td>9 (53)</td>
                <td>4 (44)</td>
                <td>.34</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>Asian or Pacific Islander</td>
                <td>1 (4)</td>
                <td>1 (6)</td>
                <td>0 (0)</td>
                <td>.34</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>Hispanic</td>
                <td>6 (23)</td>
                <td>2 (12)</td>
                <td>4 (44)</td>
                <td>.34</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>Other (eg, decline to state or multirace)</td>
                <td>2 (8)</td>
                <td>2 (12)</td>
                <td>0 (0)</td>
                <td>.34</td>
              </tr>
              <tr valign="top">
                <td colspan="6">
                  <bold>Education, n (%)</bold>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>High school or general education or less</td>
                <td>2 (8)</td>
                <td>1 (6)</td>
                <td>1 (11)</td>
                <td>.95</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>College—1 year to 3 years</td>
                <td>8 (31)</td>
                <td>5 (29)</td>
                <td>3 (33)</td>
                <td>.95</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>College—≥4 years</td>
                <td>10 (38)</td>
                <td>7 (41)</td>
                <td>3 (33)</td>
                <td>.95</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>Post college</td>
                <td>6 (23)</td>
                <td>4 (23)</td>
                <td>2 (22)</td>
                <td>.95</td>
              </tr>
              <tr valign="top">
                <td colspan="6">
                  <bold>Income (US $), n (%)</bold>
                </td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>&#60;35,000</td>
                <td>7 (27)</td>
                <td>4 (23)</td>
                <td>3 (33)</td>
                <td>.32</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>35,000 to &#60;55,000</td>
                <td>7 (27)</td>
                <td>3 (18)</td>
                <td>4 (44)</td>
                <td>.32</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>55,000 to &#60;75,000</td>
                <td>5 (19)</td>
                <td>4 (23)</td>
                <td>1 (11)</td>
                <td>.32</td>
              </tr>
              <tr valign="top">
                <td>
                  <break/>
                </td>
                <td>≥75,000</td>
                <td>7 (27)</td>
                <td>6 (35)</td>
                <td>1 (11)</td>
                <td>.32</td>
              </tr>
            </tbody>
          </table>
        </table-wrap>
      </sec>
      <sec>
        <title>User Experience, Task Load, and Working Alliance</title>
        <p>Participants had a positive evaluation (values &#62;0.8) for pragmatic (S1: mean 1.3, SD 1.2 and S2: mean 1.4, SD 0.9), hedonic (S1: mean 1.0, SD 1.1; S2: mean 1.2, SD 1.0), and overall (S1: mean 1.2, SD 1.0; S2: mean 1.3, SD 0.8) qualities related to their user experience with Lumen for both sessions. There were no statistically significant differences between the 2 sessions (t<sub>22</sub>=0.37, 0.00, and 0.25 and <italic>P</italic>=.71, .99, and .80 for pragmatic, hedonic, and overall scores, respectively).</p>
        <p>Across both sessions, participants encountered medium (approximately 50) across the mental (cognitive), effort, frustration, and performance dimensions of the NASA TLX scale. There were no statistically significant differences between S1 and S2 (<xref ref-type="table" rid="table2">Table 2</xref>). However, participants rated as having experienced more temporal workload in S2 (mean 52.0, SD 29.1) than S1 (mean 36.5, SD 23.2; <italic>P</italic>=.03), suggesting feeling rushed during their interaction with Lumen in S2.</p>
        <p>The scores on the 7-point WAI-Tech survey for task (S1: mean 5.2, SD 0.9; S2: mean 5.3, SD 0.9), bond (S1: mean 4.9, SD 1.0; S2: mean 4.7, SD 1.0), and goal (S1: mean 5.0, SD 0.9; S2: mean 5.1, SD 0.9) subscales were moderately high, indicating that Lumen-based PST sessions were perceived to be aligned with the participants’ needs, addressing their potential goals and the degree of mutual liking. There were no statistically significant differences between both sessions on the task, goal, and bond scales or the overall scores (<xref ref-type="table" rid="table3">Table 3</xref>).</p>
        <table-wrap position="float" id="table2">
          <label>Table 2</label>
          <caption>
            <p>Paired <italic>t</italic> test results comparing NASA Task Load Index scores between sessions 1 and 2.</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="430"/>
            <col width="180"/>
            <col width="180"/>
            <col width="120"/>
            <col width="90"/>
            <thead>
              <tr valign="top">
                <td>Question</td>
                <td>Session 1 (n=26), mean (SD)</td>
                <td>Session 2, (n=23), mean (SD)</td>
                <td><italic>t</italic> test (<italic>df</italic>)</td>
                <td><italic>P</italic> value</td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td>How mentally demanding was the task? (<italic>mental demand</italic><sup>a</sup>)</td>
                <td>42.7 (25.0)</td>
                <td>53.9 (26.1)</td>
                <td>−1.80 (22)</td>
                <td>.09</td>
              </tr>
              <tr valign="top">
                <td>How hurried or rushed were you in the pace of the task? (<italic>temporal demand</italic>)</td>
                <td>36.5 (23.2)</td>
                <td>52.0 (29.1)</td>
                <td>−2.37 (22)</td>
                <td>.03</td>
              </tr>
              <tr valign="top">
                <td>How hard did you have to work to accomplish your level of performance? (<italic>effort</italic>)</td>
                <td>36.0 (23.4)</td>
                <td>42.8 (18.9)</td>
                <td>−1.44 (22)</td>
                <td>.16</td>
              </tr>
              <tr valign="top">
                <td>How insecure, discouraged, irritated, stressed, and annoyed were you? (<italic>frustration</italic>)</td>
                <td>31.9 (22.0)</td>
                <td>38.5 (24.6)</td>
                <td>−0.95 (22)</td>
                <td>.35</td>
              </tr>
              <tr valign="top">
                <td>How successful were you in accomplishing what you were asked to do? (<italic>performance</italic>)</td>
                <td>34.6 (23.1)</td>
                <td>37.2 (23.3)</td>
                <td>−0.37 (22)</td>
                <td>.71</td>
              </tr>
            </tbody>
          </table>
          <table-wrap-foot>
            <fn id="table2fn1">
              <p><sup>a</sup>Italicized text shows the various categories of the NASA Task Load Index scales.</p>
            </fn>
          </table-wrap-foot>
        </table-wrap>
        <table-wrap position="float" id="table3">
          <label>Table 3</label>
          <caption>
            <p>Paired <italic>t</italic> test results comparing task, goal, and bond subscales of the Working Alliance Inventory–Technology Version between sessions 1 and 2.</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="180"/>
            <col width="260"/>
            <col width="260"/>
            <col width="180"/>
            <col width="120"/>
            <thead>
              <tr valign="top">
                <td>Scale</td>
                <td>Session 1, mean (SD)</td>
                <td>Session 2, mean (SD)</td>
                <td><italic>t</italic> test (<italic>df</italic>)</td>
                <td><italic>P</italic> value</td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td>Task subscale</td>
                <td>5.2 (0.9)</td>
                <td>5.3 (0.9)</td>
                <td>0.11 (22)</td>
                <td>.92</td>
              </tr>
              <tr valign="top">
                <td>Bond subscale</td>
                <td>4.9 (1.0)</td>
                <td>4.7 (1.0)</td>
                <td>1.49 (22)</td>
                <td>.15</td>
              </tr>
              <tr valign="top">
                <td>Goal subscale</td>
                <td>5.0 (0.9)</td>
                <td>5.1 (0.9)</td>
                <td>−0.32 (22)</td>
                <td>.75</td>
              </tr>
              <tr valign="top">
                <td>Overall scale</td>
                <td>5.0 (0.9)</td>
                <td>5.0 (0.9)</td>
                <td>0.56 (22)</td>
                <td>.58</td>
              </tr>
            </tbody>
          </table>
        </table-wrap>
      </sec>
      <sec>
        <title>User Perspectives of Lumen</title>
        <p>On the basis of the thematic analysis, we identified 6 categories that highlighted key user perspectives regarding Lumen. This included (total, N=536 coded themes across all categories; % of each category across all transcripts): (1) comparing Lumen with a human coach (ie, a human-AI comparison; 200/536, 37.3%), (2) task load experienced during Lumen interactions (102/536, 19%), (3) perception of PST delivered by Lumen (82/536, 15.2%), (4) user suggestions for improving Lumen (81/536, 15.1%), (5) natural language understanding of Lumen (44/536, 8.2%), and (6) technical issues (27/536, 5%) that were encountered during the 2 Lumen sessions (detailed descriptions of each of these categories along with exemplary quotations are provided in <xref ref-type="table" rid="table4">Table 4</xref>).</p>
        <p>Comparisons of Lumen with a human coach included several aspects: potential flexibility, ease of accessibility of Lumen for those who cannot attend face-to-face appointments, and cost-related advantages. Participants also highlighted the nonhuman nature of the interaction, describing the lack of changes in tone, emotion, instant feedback, and desiring a “more personalized human touch.” Nevertheless, nearly all participants described the potential advantages related to Lumen’s accessibility, allowing those in need for therapy easily access a coach at any time:</p>
        <disp-quote>
          <p>...the fact that the flexibility of it, the fact that I could be at home, where I could be in my car, or that, you know, I could take a moment and stop at work and go in a quiet room instead of having to, you know, actually go out and, you know, go to a building, find parking, all of the inconveniences that come with [face-to-face] appointments...</p>
        </disp-quote>
        <p>In addition, and importantly, participants with previous PST experience expressed that the Lumen sessions were similar to the human coach sessions that they had previously engaged in.</p>
        <p>Participants also highlighted the workload associated with Lumen sessions, sometimes describing the difficulty in pausing sessions to collect thoughts as they worked through the steps of PST. This was especially the case in S2, where participants were required to brainstorm multiple solutions to a problem and then list the pros and cons of each solution. The workload challenges identified were related to pacing of the sessions (temporal load) and the amount of information that was directed at the participants (cognitive load). One of the participants explained that the short time to respond made them “feel pressured to come up with something ...[...]. But she [Lumen] did ask if I needed more time, but when I was responding my answers, I [still] felt like it was a short time and I almost felt cut off<italic>.</italic>”</p>
        <p>Participants described their perceptions of the PST program or structure as well as Lumen’s role in delivering PST. Their comments highlighted the importance of the PST stepwise structured approach and Lumen’s PST coaching that enabled them to create goals that could have been overwhelming:</p>
        <disp-quote>
          <p>If my goal is truly trying and I have a problem, I just feel overwhelmed. I don’t know how to attack it. Well, Lumen supplies that. It breaks it down. It pulls all of the jumbled information out of my head, leaves the emotion behind and helps me lay out a plan for essentially attacking the problem without the emotional stress of it.</p>
        </disp-quote>
        <p>Participants provided several suggestions for improvement. This included further personalizing the PST sessions, creating embodied avatars for Lumen, incorporating a friendlier voice, and investigating ways for reducing the task load associated with the interactions. One of the most insightful aspects was several participants highlighting the importance of cognitive “offloading” [<xref ref-type="bibr" rid="ref28">28</xref>]. This was especially aligned with the need to reduce the cognitive load associated with conversational interactions, especially during the problem-solving session (S2), where participants had to identify and work through a problem, set a goal, identify and evaluate possible solutions, and then devise a structured action plan to address the problem. Participants also suggested the need for visualizing their tasks, either digital or paper-based, that would help in organizing their thought processes and saving the notes for future interactions, as highlighted in the following quote: “<italic>If it would have a way in app, I mean, [...] but like a way to help me, a way to help track for me what my progress is.</italic>”</p>
        <p>Although there were a few instances of technical issues where the participants’ verbal responses were not comprehended by Lumen because of issues related to accent or ambient noise, these issues were minimal and most users noted the ease of interaction, as described in the following quote: “I was pretty much impressed with how easy was to use and, you know, it wasn’t intimidating at all<italic>.</italic>” Additional examples of Lumen interactions including problem-solving conversations are provided in Section D in <xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref>.</p>
        <table-wrap position="float" id="table4">
          <label>Table 4</label>
          <caption>
            <p>Coding categories, their description, and examples from the interviews.</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="220"/>
            <col width="350"/>
            <col width="430"/>
            <thead>
              <tr valign="top">
                <td>Coding category (spread<sup>a</sup>, %)</td>
                <td>Description</td>
                <td>Example from data</td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td>Interactive task load (78%)</td>
                <td>Participant description of the demands of interacting with Lumen. Includes:<break/><list list-type="bullet"><list-item><p>Temporal load (pace of interactions, whether there was ample time to provide a response)</p></list-item><list-item><p>Cognitive load (density of content and length of sessions)</p></list-item></list></td>
                <td>
                  <list list-type="bullet">
                    <list-item>
                      <p>“I felt kind of rushed when it was like time to, like, think through and write things” (3502) [Temporal load]</p>
                    </list-item>
                    <list-item>
                      <p>“Sometimes it’s telling you a lot of things. So, for a user, it’s hard...You’re not looking at somebody. So, you’re really, really having to concentrate and pay attention, so if by any chance you miss something, then you kind of get lost” (1213) [Cognitive load]</p>
                    </list-item>
                  </list>
                </td>
              </tr>
              <tr valign="top">
                <td>Natural language understanding (46%)</td>
                <td>Participant description of challenges that Lumen faced with understanding participants’ verbal responses. Includes:<break/><list list-type="bullet"><list-item><p>Spoken comprehension (breakdowns due to Lumen’s comprehension)</p></list-item><list-item><p>Accent or enunciation issues (eg, understanding names)</p></list-item></list></td>
                <td>
                  <list list-type="bullet">
                    <list-item>
                      <p>“I think it was difficult to provide the prompts that were requested, and I suspect that depending on the person’s accent or if they’re from—if maybe their English isn’t exactly clear, there may be some language issues” (5457) [Spoken comprehension and accent or enunciation issues]</p>
                    </list-item>
                  </list>
                </td>
              </tr>
              <tr valign="top">
                <td>Comparison with human coach (100%)</td>
                <td>Comparison of Lumen to a human coach. Includes:<break/><list list-type="bullet"><list-item><p>Naturalness of voice or tone (presence or absence of emotion)</p></list-item><list-item><p>Interactive engagement in conversation (whether Lumen was conversational)</p></list-item><list-item><p>Lumen’s tone or inflection (identifying when Lumen was asking a question vs making a statement)</p></list-item><list-item><p>Lumen vs human PST<sup>b</sup> content (comparing depth of help Lumen provided relative to human in delivery of PST)</p></list-item><list-item><p>Perceived Lumen benefits or drawbacks (pros and cons of receiving PST from Lumen relative to human, eg, accessibility, availability, and comfort with disclosure)</p></list-item></list></td>
                <td>
                  <list list-type="bullet">
                    <list-item>
                      <p>“...just robotic. Like, I’m talking to like a machine robot. That’s my initial thought. But at the same time, not in the way that it’s like dumb, but in that it’s like very scientific and not very like human.” (6132) (Naturalness of voice or tone)</p>
                    </list-item>
                    <list-item>
                      <p>“I think initially for me, what may be missing that I picked up on right away is the human interaction component. [...] a human as opposed to talking to like a device or a computer [...] So, I don’t know how differently it'll be the more I become engaged with it.” (3498) [Interactive engagement in conversation]</p>
                    </list-item>
                    <list-item>
                      <p>“When I spoke with [the human coach], I found myself venting, if I may, and going in every which direction, whereas Lumen forces me to stay very rigid, and sometimes when going through problem solving, the emotional release of going in every which direction, direction, rather than going straight and narrow feels a lot more comfortable.” (3831) [Lumen vs human PST]</p>
                    </list-item>
                    <list-item>
                      <p>“it allows accessibility to people who can’t travel or maybe they feel anxious around talking to another person. So, it eliminates like class, it eliminates race, it eliminates sex. It eliminates sort of those prejudice that could happen in like a person-to-person to person setting.” (6132) [Perceived Lumen benefits]</p>
                    </list-item>
                  </list>
                </td>
              </tr>
              <tr valign="top">
                <td>PST features in Lumen (78%)</td>
                <td>Description of the PST features as delivered by Lumen. Includes:<break/><list list-type="bullet"><list-item><p>Program structure or format (feedback around the stepwise PST process)</p></list-item><list-item><p>Virtual PST coaching (describing Lumen’s role in the PST process)</p></list-item></list></td>
                <td>
                  <list list-type="bullet">
                    <list-item>
                      <p>“You know, I think if I’m if I am if my goal is truly trying and I have a problem, I just feel overwhelmed. I don’t know how to attack it. Well Lumen supplies that. It breaks it down. It pulls all of the jumbled information out of my head, leaves the emotion behind and helps me lay out a plan for essentially attacking the problem without the emotional stress of it.” (3831) [Program structure or format and virtual PST coaching]</p>
                    </list-item>
                  </list>
                </td>
              </tr>
              <tr valign="top">
                <td>User recommendations (62%)</td>
                <td>Participants’ recommendations for:<break/><list list-type="bullet"><list-item><p>Lumen improvements (ideas for functions or features in the user interface)</p></list-item><list-item><p>Interacting with Lumen (tips for others to have an effective session with Lumen)</p></list-item></list></td>
                <td>
                  <list list-type="bullet">
                    <list-item>
                      <p>“I would tell them that like, so like you’re talking to a computerized app, so make sure you’re speaking clearly and slowly and like follow directions in order to get what you’re what you need from it.” (6132) [Interacting with Lumen]</p>
                    </list-item>
                    <list-item>
                      <p>“I would say as a part of the app, have basically have the binder already inside the app and then maybe have a link to a principal PDF for those who want to do that.” (6023) [Lumen improvements]</p>
                    </list-item>
                    <list-item>
                      <p>“I think it would be kind of cool, especially with it being linked with Alexa is if it had the ability to pick up keywords. So, like if I, you know, saying like I need to work on my diet or trainer or whatever, that somehow it was able to tap into some of those keywords. And while it’s talking back to me saying, you know. You know, we’ve looked into like some trainings in your area. We are going to send you emails of, you know, something like that that would be like really great or hear from information regarding blah, blah, blah, blah, blah.” (3498) [Lumen improvements]</p>
                    </list-item>
                    <list-item>
                      <p>“She could be better if she if I could see it, even though is a mechanical thing or robot, I want to see Lumen, so I know how Lumen it looks...I’d rather see the person I’m talking to, even though [it] is a machine or whatever it is I would rather see, you know.” (7323) [Lumen improvements]</p>
                    </list-item>
                  </list>
                </td>
              </tr>
              <tr valign="top">
                <td>Technical issues (36%)</td>
                <td>Technical issues that were experienced by participants during the sessions. Includes:<break/><list list-type="bullet"><list-item><p>Breakdowns in conversation</p></list-item></list></td>
                <td>
                  <list list-type="bullet">
                    <list-item>
                      <p>“Well, I was a little confused when it just stopped. It was still on the app. [...] And then it just completely shut the app.” (3470) [Breakdowns in conversation]</p>
                    </list-item>
                  </list>
                </td>
              </tr>
            </tbody>
          </table>
          <table-wrap-foot>
            <fn id="table4fn1">
              <p><sup>a</sup>Spread refers to the percentage of transcripts (total=50) that the coding category was present.</p>
            </fn>
            <fn id="table4fn2">
              <p><sup>b</sup>PST: problem-solving treatment.</p>
            </fn>
          </table-wrap-foot>
        </table-wrap>
      </sec>
    </sec>
    <sec sec-type="discussion">
      <title>Discussion</title>
      <sec>
        <title>Principal Findings</title>
        <p>We designed and developed a virtual voice-based coach, Lumen, which delivers an evidence-based PST program for depression and anxiety. To the best of our knowledge, Lumen is one of the first voice-based virtual coach application for delivering behavioral therapy. In contrast to prior research that has primarily used voice assistants in web-based information–seeking tasks, Lumen delivers therapy aligned with the goals and principles of an empirically validated PST program. In this developmental evaluation, participants found the Lumen virtual coach to have high pragmatic usability and user experience, with limited task load during interactions. Participants also highlighted the considerable advantages of Lumen including the on-demand accessibility to a virtual therapist and the delivery of a complex PST task with a simplistic structure and organization for achieving therapy goals. Moreover, although the second session required increased user input, there were no marked differences in effort or interaction quality, except for temporal load (associated with the pace of the conversations), which was highlighted by the participants in their interviews. In addition, the participants highlighted the lack of personalization and deep engagement in the conversation and the relative lack of emotional engagement in the conversations.</p>
      </sec>
      <sec>
        <title>Comparison With Prior Work</title>
        <p>PST, traditionally delivered by human coaches in face-to-face or phone-based settings, has been developed on mobile platforms [<xref ref-type="bibr" rid="ref29">29</xref>]. However, similar to other text-based mobile apps, participant engagement with mobile PST platforms has been challenging [<xref ref-type="bibr" rid="ref30">30</xref>]. To this end, Lumen offers a novel, voice-based mechanism for seemingly naturalistic voice interactions, potentially replicating interactions with a therapist. As previously described, much of the prior work has relied on evaluating the quality health information–seeking tasks using voice-based personal assistants (eg, [<xref ref-type="bibr" rid="ref8">8</xref>,<xref ref-type="bibr" rid="ref9">9</xref>]). Moreover, many of the previously developed applications have been preliminary prototypes (eg, [<xref ref-type="bibr" rid="ref15">15</xref>]) that lacked extensive evaluation or outcome assessment. To the best of our knowledge, this is one of the first fully functional voice-based applications that provides end-to-end support for behavioral therapy (in this case, PST).</p>
      </sec>
      <sec>
        <title>Design Changes</title>
        <p>Several design changes were incorporated in response to participants’ suggestions. To reduce the temporal and cognitive load (ie, reducing the pace of conversations), we incorporated multiple functionalities within Lumen. First, we split longer conversations (especially in S1, where Lumen provided an overview of the PST) into multiple shorter conversations to reduce the mean length of conversations between Lumen and the participant. Such shorter conversations allow for more interactive turns and have been shown to improve the common ground and engagement between conversational partners [<xref ref-type="bibr" rid="ref5">5</xref>,<xref ref-type="bibr" rid="ref31">31</xref>-<xref ref-type="bibr" rid="ref34">34</xref>]. Second, we developed functionality that allowed participants to repeat, pause, and resume conversations. This allowed participants to ask Lumen to repeat instructions if they could not keep up with the content or to pause conversations in situations where they needed to take a break. Finally, we slowed the pace of the conversations to reduce temporal demand.</p>
        <p>In addition, based on suggestions, we also developed a workbook to accompany Lumen in both physical and digital forms. The workbook includes content corresponding to the PST and simple worksheets for taking notes and facilitating brainstorming problem-solving goals, developing and evaluating potential solutions, and creating action plans. Such a cognitive aid helps in externalizing the thought processes [<xref ref-type="bibr" rid="ref28">28</xref>,<xref ref-type="bibr" rid="ref35">35</xref>,<xref ref-type="bibr" rid="ref36">36</xref>] and creating a record for follow-up after the session. Recording and brainstorming with tools also affords cognitive benefits, especially with older adults, such as prospective memory regarding the goals and action plans that were created [<xref ref-type="bibr" rid="ref37">37</xref>]. We also developed several features linked to Lumen to further integrate contextual aspects regarding the user including their current status and progress. For example, participants can track their progress by viewing their completed sessions and responses to the Patient Health Questionnaire-9 and General Anxiety Disorder-7 surveys on a user dashboard. Similarly, responses on the Patient Health Questionnaire-9 and General Anxiety Disorder-7 surveys were integrated into the Lumen session and reviewed during the session to help participants monitor the level of their depressive and anxiety symptoms.</p>
        <p>Finally, we heeded several privacy and security considerations for pragmatic implementation and testing in a real-world setting within the context of a planned pilot randomized clinical trial. To this end, we will afford trial participants access to the Lumen skill within the Amazon Alexa app on a fully encrypted and locked down iPad, with timed exits for nonuse. This allows for preventing accidental recording issues that have been reported regarding the use of voice-based smart devices (Section B, <xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref>). The iPad-based delivery is aligned with the concept of using a stand-alone “device as a therapist” for the planned trial. However, additional considerations regarding voice-based profile verifications and security considerations are necessary for a wider, pragmatic use of Lumen as a daily therapy tool.</p>
        <p>Despite these changes, several aspects of Lumen’s design and interaction are limited by current AI-based voice technology. In particular, the natural language understanding challenges of voice-based technology are well documented [<xref ref-type="bibr" rid="ref10">10</xref>]. These include difficulties in parsing tone, accent, and pronunciation in spoken language, creating breakdowns in conversation and making it functionally impossible to have a free-form, open-ended conversation with these devices. In addition, current technology is also not able to discern differences in emotion or other verbal cues that are easily interpreted in face-to-face human conversations [<xref ref-type="bibr" rid="ref5">5</xref>]. With ongoing improvements in technology, these challenges are likely to be mitigated over time, allowing for continued improvement of Lumen for optimized user experience.</p>
      </sec>
      <sec>
        <title>Limitations</title>
        <p>This mixed methods formative research study had several limitations. The study was based on a small sample of users (N=26) who used Lumen in a relatively controlled environment. However, participants were engaged in 2 sessions and performed the Lumen interactions without external support. Only 2 sessions were evaluated with participants, and as such, we could not characterize participants’ experience with the entire 8-session PST program. However, structurally, sessions 2 to 8 mirror the S2 evaluated in this study. It is likely that participants will become more or less comfortable with the Lumen interactions in the later sessions. Given the formative and controlled nature of this study, we could not assess the impact of the various measures (ie, task load and work alliance) over time. We will be able to determine such longitudinal effects in our ongoing pilot clinical trial. Sessions were attended by a research coordinator and a trained note taker. It is not known whether their presence influenced the participants’ use of Lumen or their responses to the interview questions.</p>
        <p>Notwithstanding those technological and research limitations, the findings from the formative evaluation and the subsequent improvements in design and functionalities position Lumen to be a “minimum viable product” that is highly acceptable to participants, appears to veridically reflect PST content, and is ready for potential real-world pilot testing. Recruitment has been completed for the pilot clinical trial (ClinicalTrials.gov, NCT# 04524104) in which 63 adults with mild to moderate depressive and anxiety symptoms have been randomized in a 2:1 ratio to the Lumen intervention or the wait-list control group and followed for 4 months. The objectives of the pilot trial are 3-fold: (1) to determine the feasibility and acceptability of the Lumen virtual coach for delivering the 8-session PST program; (2) to assess neural target engagement by comparing changes in the amygdala and dorsal lateral prefrontal cortex in functional neuroimaging between the Lumen intervention and wait-list control groups; and (3) to examine the relationship between neural target engagement and changes in self-reported measures of mood, coping, and psychosocial functioning. The pilot trial will provide the preliminary data needed to accelerate the clinical and translational research on this novel digital psychotherapy and to catalyze future development and definitive efficacy clinical trials.</p>
      </sec>
      <sec>
        <title>Conclusions</title>
        <p>With a goal of overcoming the lack of empirical evidence for AI-based voice applications in behavioral therapy, we developed a voice-only virtual coach, Lumen, for delivering PST. The findings from the formative evaluation highlight feasibility, accessibility, and favorable user experience. Suggestions for more natural conversations and better contextual support have resulted in an improved, minimally viable product. Lumen is being tested in a clinical trial to evaluate its neural mechanism of action and therapeutic potential in depression and anxiety. If successful, Lumen can be a viable voice-based therapist offering a realistic and cognitively plausible verbal interaction for personalized and accessible mental health care, filling a gap in traditional mental health services.</p>
      </sec>
    </sec>
  </body>
  <back>
    <app-group>
      <supplementary-material id="app1">
        <label>Multimedia Appendix 1</label>
        <p>Details of the Lumen architecture and user interaction patterns with Lumen.</p>
        <media xlink:href="formative_v6i8e38092_app1.docx" xlink:title="DOCX File , 456 KB"/>
      </supplementary-material>
    </app-group>
    <glossary>
      <title>Abbreviations</title>
      <def-list>
        <def-item>
          <term id="abb1">AI</term>
          <def>
            <p>artificial intelligence</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb2">ENGAGE-2</term>
          <def>
            <p>Engaging Self-Regulation Targets to Understand the Mechanisms of Behavior Change and Improve Mood and Weight Outcomes</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb3">PST</term>
          <def>
            <p>problem-solving treatment</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb4">TLX</term>
          <def>
            <p>Task Load Index</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb5">UEQ-S</term>
          <def>
            <p>User Experience Questionnaire Short Version</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb6">WAI-Tech</term>
          <def>
            <p>Working Alliance Inventory–Technology Version</p>
          </def>
        </def-item>
      </def-list>
    </glossary>
    <ack>
      <p>This study was funded by a grant from the National Institute of Mental Health (grant number R61MH119237).</p>
    </ack>
    <notes>
      <sec>
        <title>Data Availability</title>
        <p>Deidentified data from this study are not available in a public archive. Deidentified data from this study will be made available (as permissible according to institutional review board standards) by emailing the corresponding author.</p>
      </sec>
    </notes>
    <fn-group>
      <fn fn-type="con">
        <p>TK, CRR, NEW, JMS, and JM conceived the study; TK, CRR, EAK, and NL collected the data; TK, CRR, EAK, NL, and NEW were involved in the preliminary analysis; and all authors participated in the interpretation of the results, writing of the manuscript, and critical review. All authors approved the final manuscript for submission.</p>
      </fn>
      <fn fn-type="conflict">
        <p>TK is a paid consultant for Pfizer, Inc outside of this work. JM is a paid scientific consultant for Health Mentor Inc (San Jose, California, United States). OAA is the cofounder of KeyWise AI and serves on the advisory boards of Blueprint Health and Embodied Labs. The other authors report no conflicts of interest.</p>
      </fn>
    </fn-group>
    <ref-list>
      <ref id="ref1">
        <label>1</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Topol</surname>
              <given-names>EJ</given-names>
            </name>
          </person-group>
          <article-title>High-performance medicine: the convergence of human and artificial intelligence</article-title>
          <source>Nat Med</source>
          <year>2019</year>
          <month>01</month>
          <day>7</day>
          <volume>25</volume>
          <issue>1</issue>
          <fpage>44</fpage>
          <lpage>56</lpage>
          <pub-id pub-id-type="doi">10.1038/s41591-018-0300-7</pub-id>
          <pub-id pub-id-type="medline">30617339</pub-id>
          <pub-id pub-id-type="pii">10.1038/s41591-018-0300-7</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref2">
        <label>2</label>
        <nlm-citation citation-type="web">
          <article-title>New data on voice assistant SEO is a wake-up call for brands</article-title>
          <source>voicebot.ai</source>
          <access-date>2021-04-01</access-date>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://voicebot.ai/2019/07/09/new-data-on-voice-assistant-seo-is-a-wake-up-call-for-brands/">https://voicebot.ai/2019/07/09/new-data-on -voice-assistant-seo-is-a-wake-up-call-for-brands/</ext-link>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref3">
        <label>3</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Steinhubl</surname>
              <given-names>SR</given-names>
            </name>
            <name name-style="western">
              <surname>Topol</surname>
              <given-names>EJ</given-names>
            </name>
          </person-group>
          <article-title>Now we're talking: bringing a voice to digital medicine</article-title>
          <source>Lancet</source>
          <year>2018</year>
          <month>08</month>
          <volume>392</volume>
          <issue>10148</issue>
          <fpage>627</fpage>
          <pub-id pub-id-type="doi">10.1016/s0140-6736(18)31803-8</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref4">
        <label>4</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Sezgin</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Militello</surname>
              <given-names>LK</given-names>
            </name>
            <name name-style="western">
              <surname>Huang</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Lin</surname>
              <given-names>S</given-names>
            </name>
          </person-group>
          <article-title>A scoping review of patient-facing, behavioral health interventions with voice assistant technology targeting self-management and healthy lifestyle behaviors</article-title>
          <source>Transl Behav Med</source>
          <year>2020</year>
          <month>08</month>
          <day>07</day>
          <volume>10</volume>
          <issue>3</issue>
          <fpage>606</fpage>
          <lpage>28</lpage>
          <pub-id pub-id-type="doi">10.1093/tbm/ibz141</pub-id>
          <pub-id pub-id-type="medline">32766865</pub-id>
          <pub-id pub-id-type="pii">5885015</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref5">
        <label>5</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Kannampallil</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Smyth</surname>
              <given-names>JM</given-names>
            </name>
            <name name-style="western">
              <surname>Jones</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Payne</surname>
              <given-names>PR</given-names>
            </name>
            <name name-style="western">
              <surname>Ma</surname>
              <given-names>J</given-names>
            </name>
          </person-group>
          <article-title>Cognitive plausibility in voice-based AI health counselors</article-title>
          <source>NPJ Digit Med</source>
          <year>2020</year>
          <month>05</month>
          <day>15</day>
          <volume>3</volume>
          <issue>1</issue>
          <fpage>72</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.1038/s41746-020-0278-7"/>
          </comment>
          <pub-id pub-id-type="doi">10.1038/s41746-020-0278-7</pub-id>
          <pub-id pub-id-type="medline">32435700</pub-id>
          <pub-id pub-id-type="pii">278</pub-id>
          <pub-id pub-id-type="pmcid">PMC7229176</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref6">
        <label>6</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Boyd</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Wilson</surname>
              <given-names>N</given-names>
            </name>
          </person-group>
          <article-title>Just ask Siri? A pilot study comparing smartphone digital assistants and laptop Google searches for smoking cessation advice</article-title>
          <source>PLoS One</source>
          <year>2018</year>
          <month>3</month>
          <day>28</day>
          <volume>13</volume>
          <issue>3</issue>
          <fpage>e0194811</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://dx.plos.org/10.1371/journal.pone.0194811"/>
          </comment>
          <pub-id pub-id-type="doi">10.1371/journal.pone.0194811</pub-id>
          <pub-id pub-id-type="medline">29590168</pub-id>
          <pub-id pub-id-type="pii">PONE-D-17-42760</pub-id>
          <pub-id pub-id-type="pmcid">PMC5874038</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref7">
        <label>7</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Kocaballi</surname>
              <given-names>AB</given-names>
            </name>
            <name name-style="western">
              <surname>Quiroz</surname>
              <given-names>JC</given-names>
            </name>
            <name name-style="western">
              <surname>Rezazadegan</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Berkovsky</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Magrabi</surname>
              <given-names>F</given-names>
            </name>
            <name name-style="western">
              <surname>Coiera</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Laranjo</surname>
              <given-names>L</given-names>
            </name>
          </person-group>
          <article-title>Responses of conversational agents to health and lifestyle prompts: investigation of appropriateness and presentation structures</article-title>
          <source>J Med Internet Res</source>
          <year>2020</year>
          <month>02</month>
          <day>09</day>
          <volume>22</volume>
          <issue>2</issue>
          <fpage>e15823</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.jmir.org/2020/2/e15823/"/>
          </comment>
          <pub-id pub-id-type="doi">10.2196/15823</pub-id>
          <pub-id pub-id-type="medline">32039810</pub-id>
          <pub-id pub-id-type="pii">v22i2e15823</pub-id>
          <pub-id pub-id-type="pmcid">PMC7055771</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref8">
        <label>8</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Miner</surname>
              <given-names>AS</given-names>
            </name>
            <name name-style="western">
              <surname>Milstein</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Schueller</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Hegde</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Mangurian</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Linos</surname>
              <given-names>E</given-names>
            </name>
          </person-group>
          <article-title>Smartphone-based conversational agents and responses to questions about mental health, interpersonal violence, and physical health</article-title>
          <source>JAMA Intern Med</source>
          <year>2016</year>
          <month>05</month>
          <day>01</day>
          <volume>176</volume>
          <issue>5</issue>
          <fpage>619</fpage>
          <lpage>25</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="http://europepmc.org/abstract/MED/26974260"/>
          </comment>
          <pub-id pub-id-type="doi">10.1001/jamainternmed.2016.0400</pub-id>
          <pub-id pub-id-type="medline">26974260</pub-id>
          <pub-id pub-id-type="pii">2500043</pub-id>
          <pub-id pub-id-type="pmcid">PMC4996669</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref9">
        <label>9</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Nobles</surname>
              <given-names>AL</given-names>
            </name>
            <name name-style="western">
              <surname>Leas</surname>
              <given-names>EC</given-names>
            </name>
            <name name-style="western">
              <surname>Caputi</surname>
              <given-names>TL</given-names>
            </name>
            <name name-style="western">
              <surname>Zhu</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Strathdee</surname>
              <given-names>SA</given-names>
            </name>
            <name name-style="western">
              <surname>Ayers</surname>
              <given-names>JW</given-names>
            </name>
          </person-group>
          <article-title>Responses to addiction help-seeking from Alexa, Siri, Google Assistant, Cortana, and Bixby intelligent virtual assistants</article-title>
          <source>NPJ Digit Med</source>
          <year>2020</year>
          <month>01</month>
          <day>29</day>
          <volume>3</volume>
          <issue>1</issue>
          <fpage>11</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.1038/s41746-019-0215-9"/>
          </comment>
          <pub-id pub-id-type="doi">10.1038/s41746-019-0215-9</pub-id>
          <pub-id pub-id-type="medline">32025572</pub-id>
          <pub-id pub-id-type="pii">215</pub-id>
          <pub-id pub-id-type="pmcid">PMC6989668</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref10">
        <label>10</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Bickmore</surname>
              <given-names>TW</given-names>
            </name>
            <name name-style="western">
              <surname>Trinh</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Olafsson</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>O'Leary</surname>
              <given-names>TK</given-names>
            </name>
            <name name-style="western">
              <surname>Asadi</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Rickles</surname>
              <given-names>NM</given-names>
            </name>
            <name name-style="western">
              <surname>Cruz</surname>
              <given-names>R</given-names>
            </name>
          </person-group>
          <article-title>Patient and consumer safety risks when using conversational assistants for medical information: an observational study of siri, alexa, and google assistant</article-title>
          <source>J Med Internet Res</source>
          <year>2018</year>
          <month>09</month>
          <day>04</day>
          <volume>20</volume>
          <issue>9</issue>
          <fpage>e11510</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.jmir.org/2018/9/e11510/"/>
          </comment>
          <pub-id pub-id-type="doi">10.2196/11510</pub-id>
          <pub-id pub-id-type="medline">30181110</pub-id>
          <pub-id pub-id-type="pii">v20i9e11510</pub-id>
          <pub-id pub-id-type="pmcid">PMC6231817</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref11">
        <label>11</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Alagha</surname>
              <given-names>EC</given-names>
            </name>
            <name name-style="western">
              <surname>Helbing</surname>
              <given-names>RR</given-names>
            </name>
          </person-group>
          <article-title>Evaluating the quality of voice assistants' responses to consumer health questions about vaccines: an exploratory comparison of Alexa, Google Assistant and Siri</article-title>
          <source>BMJ Health Care Inform</source>
          <year>2019</year>
          <month>11</month>
          <day>24</day>
          <volume>26</volume>
          <issue>1</issue>
          <fpage>e100075</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://informatics.bmj.com/lookup/pmidlookup?view=long&#38;pmid=31767629"/>
          </comment>
          <pub-id pub-id-type="doi">10.1136/bmjhci-2019-100075</pub-id>
          <pub-id pub-id-type="medline">31767629</pub-id>
          <pub-id pub-id-type="pii">bmjhci-2019-100075</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref12">
        <label>12</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Yang</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Lee</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Sezgin</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Bridge</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Lin</surname>
              <given-names>S</given-names>
            </name>
          </person-group>
          <article-title>Clinical advice by voice assistants on postpartum depression: cross-sectional investigation using apple Siri, Amazon Alexa, Google Assistant, and Microsoft Cortana</article-title>
          <source>JMIR Mhealth Uhealth</source>
          <year>2021</year>
          <month>01</month>
          <day>11</day>
          <volume>9</volume>
          <issue>1</issue>
          <fpage>e24045</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://mhealth.jmir.org/2021/1/e24045/"/>
          </comment>
          <pub-id pub-id-type="doi">10.2196/24045</pub-id>
          <pub-id pub-id-type="medline">33427680</pub-id>
          <pub-id pub-id-type="pii">v9i1e24045</pub-id>
          <pub-id pub-id-type="pmcid">PMC7834933</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref13">
        <label>13</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Palanica</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Thommandram</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Lee</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Li</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Fossat</surname>
              <given-names>Y</given-names>
            </name>
          </person-group>
          <article-title>Do you understand the words that are comin outta my mouth? Voice assistant comprehension of medication names</article-title>
          <source>NPJ Digit Med</source>
          <year>2019</year>
          <month>6</month>
          <day>20</day>
          <volume>2</volume>
          <issue>1</issue>
          <fpage>55</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.1038/s41746-019-0133-x"/>
          </comment>
          <pub-id pub-id-type="doi">10.1038/s41746-019-0133-x</pub-id>
          <pub-id pub-id-type="medline">31304401</pub-id>
          <pub-id pub-id-type="pii">133</pub-id>
          <pub-id pub-id-type="pmcid">PMC6586879</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref14">
        <label>14</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Wilson</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>MacDonald</surname>
              <given-names>EJ</given-names>
            </name>
            <name name-style="western">
              <surname>Mansoor</surname>
              <given-names>OD</given-names>
            </name>
            <name name-style="western">
              <surname>Morgan</surname>
              <given-names>J</given-names>
            </name>
          </person-group>
          <article-title>In bed with Siri and Google Assistant: a comparison of sexual health advice</article-title>
          <source>BMJ</source>
          <year>2017</year>
          <month>12</month>
          <day>13</day>
          <volume>359</volume>
          <fpage>j5635</fpage>
          <pub-id pub-id-type="doi">10.1136/bmj.j5635</pub-id>
          <pub-id pub-id-type="medline">29237603</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref15">
        <label>15</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Ismail</surname>
              <given-names>HO</given-names>
            </name>
            <name name-style="western">
              <surname>Moses</surname>
              <given-names>AR</given-names>
            </name>
            <name name-style="western">
              <surname>Tadrus</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Mohamed</surname>
              <given-names>EA</given-names>
            </name>
            <name name-style="western">
              <surname>Jones</surname>
              <given-names>LS</given-names>
            </name>
          </person-group>
          <article-title>Feasibility of use of a smart speaker to administer Snellen visual acuity examinations in a clinical setting</article-title>
          <source>JAMA Netw Open</source>
          <year>2020</year>
          <month>08</month>
          <day>03</day>
          <volume>3</volume>
          <issue>8</issue>
          <fpage>e2013908</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://jamanetwork.com/journals/jamanetworkopen/fullarticle/10.1001/jamanetworkopen.2020.13908"/>
          </comment>
          <pub-id pub-id-type="doi">10.1001/jamanetworkopen.2020.13908</pub-id>
          <pub-id pub-id-type="medline">32822489</pub-id>
          <pub-id pub-id-type="pii">2769502</pub-id>
          <pub-id pub-id-type="pmcid">PMC7439105</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref16">
        <label>16</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Cheng</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Raghavaraju</surname>
              <given-names>V</given-names>
            </name>
            <name name-style="western">
              <surname>Kanugo</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Handrianto</surname>
              <given-names>YP</given-names>
            </name>
            <name name-style="western">
              <surname>Shang</surname>
              <given-names>Y</given-names>
            </name>
          </person-group>
          <article-title>Development and evaluation of a healthy coping voice interface application using the Google home for elderly patients with type 2 diabetes</article-title>
          <source>Proceedings of the 2018 15th IEEE Annual Consumer Communications &#38; Networking Conference (CCNC)</source>
          <year>2018</year>
          <conf-name>2018 15th IEEE Annual Consumer Communications &#38; Networking Conference (CCNC)</conf-name>
          <conf-date>Jan 12-15, 2018</conf-date>
          <conf-loc>Las Vegas, NV, USA</conf-loc>
          <pub-id pub-id-type="doi">10.1109/ccnc.2018.8319283</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref17">
        <label>17</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Li</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Maharjan</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Xie</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Tao</surname>
              <given-names>C</given-names>
            </name>
          </person-group>
          <article-title>A personalized voice-based diet assistant for caregivers of Alzheimer disease and related dementias: system development and validation</article-title>
          <source>J Med Internet Res</source>
          <year>2020</year>
          <month>09</month>
          <day>21</day>
          <volume>22</volume>
          <issue>9</issue>
          <fpage>e19897</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.jmir.org/2020/9/e19897/"/>
          </comment>
          <pub-id pub-id-type="doi">10.2196/19897</pub-id>
          <pub-id pub-id-type="medline">32955452</pub-id>
          <pub-id pub-id-type="pii">v22i9e19897</pub-id>
          <pub-id pub-id-type="pmcid">PMC7536606</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref18">
        <label>18</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Vaidyam</surname>
              <given-names>AN</given-names>
            </name>
            <name name-style="western">
              <surname>Wisniewski</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Halamka</surname>
              <given-names>JD</given-names>
            </name>
            <name name-style="western">
              <surname>Kashavan</surname>
              <given-names>MS</given-names>
            </name>
            <name name-style="western">
              <surname>Torous</surname>
              <given-names>JB</given-names>
            </name>
          </person-group>
          <article-title>Chatbots and conversational agents in mental health: a review of the psychiatric landscape</article-title>
          <source>Can J Psychiatry</source>
          <year>2019</year>
          <month>07</month>
          <day>21</day>
          <volume>64</volume>
          <issue>7</issue>
          <fpage>456</fpage>
          <lpage>64</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="http://europepmc.org/abstract/MED/30897957"/>
          </comment>
          <pub-id pub-id-type="doi">10.1177/0706743719828977</pub-id>
          <pub-id pub-id-type="medline">30897957</pub-id>
          <pub-id pub-id-type="pmcid">PMC6610568</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref19">
        <label>19</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Pham</surname>
              <given-names>KT</given-names>
            </name>
            <name name-style="western">
              <surname>Nabizadeh</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Selek</surname>
              <given-names>S</given-names>
            </name>
          </person-group>
          <article-title>Artificial intelligence and chatbots in psychiatry</article-title>
          <source>Psychiatr Q</source>
          <year>2022</year>
          <month>03</month>
          <day>25</day>
          <volume>93</volume>
          <issue>1</issue>
          <fpage>249</fpage>
          <lpage>53</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="http://europepmc.org/abstract/MED/35212940"/>
          </comment>
          <pub-id pub-id-type="doi">10.1007/s11126-022-09973-8</pub-id>
          <pub-id pub-id-type="medline">35212940</pub-id>
          <pub-id pub-id-type="pii">10.1007/s11126-022-09973-8</pub-id>
          <pub-id pub-id-type="pmcid">PMC8873348</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref20">
        <label>20</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Fitzpatrick</surname>
              <given-names>KK</given-names>
            </name>
            <name name-style="western">
              <surname>Darcy</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Vierhile</surname>
              <given-names>M</given-names>
            </name>
          </person-group>
          <article-title>Delivering cognitive behavior therapy to young adults with symptoms of depression and anxiety using a fully automated conversational agent (Woebot): a randomized controlled trial</article-title>
          <source>JMIR Ment Health</source>
          <year>2017</year>
          <month>06</month>
          <day>06</day>
          <volume>4</volume>
          <issue>2</issue>
          <fpage>e19</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://mental.jmir.org/2017/2/e19/"/>
          </comment>
          <pub-id pub-id-type="doi">10.2196/mental.7785</pub-id>
          <pub-id pub-id-type="medline">28588005</pub-id>
          <pub-id pub-id-type="pii">v4i2e19</pub-id>
          <pub-id pub-id-type="pmcid">PMC5478797</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref21">
        <label>21</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Ma</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Rosas</surname>
              <given-names>LG</given-names>
            </name>
            <name name-style="western">
              <surname>Lv</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Xiao</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Snowden</surname>
              <given-names>MB</given-names>
            </name>
            <name name-style="western">
              <surname>Venditti</surname>
              <given-names>EM</given-names>
            </name>
            <name name-style="western">
              <surname>Lewis</surname>
              <given-names>MA</given-names>
            </name>
            <name name-style="western">
              <surname>Goldhaber-Fiebert</surname>
              <given-names>JD</given-names>
            </name>
            <name name-style="western">
              <surname>Lavori</surname>
              <given-names>PW</given-names>
            </name>
          </person-group>
          <article-title>Effect of integrated behavioral weight loss treatment and problem-solving therapy on body mass index and depressive symptoms among patients with obesity and depression: the RAINBOW randomized clinical trial</article-title>
          <source>JAMA</source>
          <year>2019</year>
          <month>03</month>
          <day>05</day>
          <volume>321</volume>
          <issue>9</issue>
          <fpage>869</fpage>
          <lpage>79</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="http://europepmc.org/abstract/MED/30835308"/>
          </comment>
          <pub-id pub-id-type="doi">10.1001/jama.2019.0557</pub-id>
          <pub-id pub-id-type="medline">30835308</pub-id>
          <pub-id pub-id-type="pii">2726984</pub-id>
          <pub-id pub-id-type="pmcid">PMC6439596</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref22">
        <label>22</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Schrepp</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Hinderks</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Thomaschewski</surname>
              <given-names>J</given-names>
            </name>
          </person-group>
          <article-title>Design and evaluation of a short version of the User Experience Questionnaire (UEQ-S)</article-title>
          <source>Int J Interact Multimedia Artif Intell</source>
          <year>2017</year>
          <volume>4</volume>
          <issue>6</issue>
          <fpage>103</fpage>
          <pub-id pub-id-type="doi">10.9781/ijimai.2017.09.001</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref23">
        <label>23</label>
        <nlm-citation citation-type="book">
          <article-title>Development of NASA-TLX (Task Load Index): results of empirical and theoretical research</article-title>
          <source>Advances in Psychology</source>
          <year>1988</year>
          <publisher-loc>Amsterdam, Netherlands</publisher-loc>
          <publisher-name>Elsevier Science</publisher-name>
        </nlm-citation>
      </ref>
      <ref id="ref24">
        <label>24</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Kiluk</surname>
              <given-names>BD</given-names>
            </name>
            <name name-style="western">
              <surname>Serafini</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Frankforter</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Nich</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Carroll</surname>
              <given-names>KM</given-names>
            </name>
          </person-group>
          <article-title>Only connect: the working alliance in computer-based cognitive behavioral therapy</article-title>
          <source>Behav Res Ther</source>
          <year>2014</year>
          <month>12</month>
          <volume>63</volume>
          <fpage>139</fpage>
          <lpage>46</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="http://europepmc.org/abstract/MED/25461789"/>
          </comment>
          <pub-id pub-id-type="doi">10.1016/j.brat.2014.10.003</pub-id>
          <pub-id pub-id-type="medline">25461789</pub-id>
          <pub-id pub-id-type="pii">S0005-7967(14)00166-1</pub-id>
          <pub-id pub-id-type="pmcid">PMC4408209</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref25">
        <label>25</label>
        <nlm-citation citation-type="book">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Clarke</surname>
              <given-names>V</given-names>
            </name>
            <name name-style="western">
              <surname>Braun</surname>
              <given-names>V</given-names>
            </name>
          </person-group>
          <article-title>Thematic analysis</article-title>
          <source>Encyclopedia of Critical Psychology</source>
          <year>2014</year>
          <publisher-loc>New York</publisher-loc>
          <publisher-name>Springer</publisher-name>
        </nlm-citation>
      </ref>
      <ref id="ref26">
        <label>26</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Crowe</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Inder</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Porter</surname>
              <given-names>R</given-names>
            </name>
          </person-group>
          <article-title>Conducting qualitative research in mental health: thematic and content analyses</article-title>
          <source>Aust N Z J Psychiatry</source>
          <year>2015</year>
          <month>07</month>
          <day>21</day>
          <volume>49</volume>
          <issue>7</issue>
          <fpage>616</fpage>
          <lpage>23</lpage>
          <pub-id pub-id-type="doi">10.1177/0004867415582053</pub-id>
          <pub-id pub-id-type="medline">25900973</pub-id>
          <pub-id pub-id-type="pii">0004867415582053</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref27">
        <label>27</label>
        <nlm-citation citation-type="web">
          <article-title>User Experience Questionnaire</article-title>
          <source>UEQ</source>
          <access-date>2021-04-01</access-date>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.ueq-online.org/">https://www.ueq-online.org/</ext-link>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref28">
        <label>28</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Larkin</surname>
              <given-names>PH</given-names>
            </name>
            <name name-style="western">
              <surname>Simon</surname>
              <given-names>HA</given-names>
            </name>
          </person-group>
          <article-title>Why a diagram is (sometimes) worth ten thousand words</article-title>
          <source>Cognit Sci Multidisciplin J</source>
          <year>1987</year>
          <volume>11</volume>
          <issue>1</issue>
          <fpage>65</fpage>
          <lpage>100</lpage>
          <pub-id pub-id-type="doi">10.1111/j.1551-6708.1987.tb00863.x</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref29">
        <label>29</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Arean</surname>
              <given-names>PA</given-names>
            </name>
            <name name-style="western">
              <surname>Hallgren</surname>
              <given-names>KA</given-names>
            </name>
            <name name-style="western">
              <surname>Jordan</surname>
              <given-names>JT</given-names>
            </name>
            <name name-style="western">
              <surname>Gazzaley</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Atkins</surname>
              <given-names>DC</given-names>
            </name>
            <name name-style="western">
              <surname>Heagerty</surname>
              <given-names>PJ</given-names>
            </name>
            <name name-style="western">
              <surname>Anguera</surname>
              <given-names>JA</given-names>
            </name>
          </person-group>
          <article-title>The use and effectiveness of mobile apps for depression: results from a fully remote clinical trial</article-title>
          <source>J Med Internet Res</source>
          <year>2016</year>
          <month>12</month>
          <day>20</day>
          <volume>18</volume>
          <issue>12</issue>
          <fpage>e330</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.jmir.org/2016/12/e330/"/>
          </comment>
          <pub-id pub-id-type="doi">10.2196/jmir.6482</pub-id>
          <pub-id pub-id-type="medline">27998876</pub-id>
          <pub-id pub-id-type="pii">v18i12e330</pub-id>
          <pub-id pub-id-type="pmcid">PMC5209607</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref30">
        <label>30</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Pratap</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Renn</surname>
              <given-names>BN</given-names>
            </name>
            <name name-style="western">
              <surname>Volponi</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Mooney</surname>
              <given-names>SD</given-names>
            </name>
            <name name-style="western">
              <surname>Gazzaley</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Arean</surname>
              <given-names>PA</given-names>
            </name>
            <name name-style="western">
              <surname>Anguera</surname>
              <given-names>JA</given-names>
            </name>
          </person-group>
          <article-title>Using mobile apps to assess and treat depression in Hispanic and Latino populations: fully remote randomized clinical trial</article-title>
          <source>J Med Internet Res</source>
          <year>2018</year>
          <month>08</month>
          <day>09</day>
          <volume>20</volume>
          <issue>8</issue>
          <fpage>e10130</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.jmir.org/2018/8/e10130/"/>
          </comment>
          <pub-id pub-id-type="doi">10.2196/10130</pub-id>
          <pub-id pub-id-type="medline">30093372</pub-id>
          <pub-id pub-id-type="pii">v20i8e10130</pub-id>
          <pub-id pub-id-type="pmcid">PMC6107735</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref31">
        <label>31</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Brennan</surname>
              <given-names>SE</given-names>
            </name>
            <name name-style="western">
              <surname>Clark</surname>
              <given-names>HH</given-names>
            </name>
          </person-group>
          <article-title>Conceptual pacts and lexical choice in conversation</article-title>
          <source>J Exp Psychol Learn Mem Cogn</source>
          <year>1996</year>
          <volume>22</volume>
          <issue>6</issue>
          <fpage>1482</fpage>
          <lpage>93</lpage>
          <pub-id pub-id-type="doi">10.1037/0278-7393.22.6.1482</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref32">
        <label>32</label>
        <nlm-citation citation-type="book">
          <article-title>Two minds, one dialog: coordinating speaking and understanding</article-title>
          <source>Psychology of Learning and Motivation</source>
          <year>2010</year>
          <publisher-loc>Amsterdam, Netherlands</publisher-loc>
          <publisher-name>Elsevier Science</publisher-name>
        </nlm-citation>
      </ref>
      <ref id="ref33">
        <label>33</label>
        <nlm-citation citation-type="book">
          <source>Using Language</source>
          <year>1996</year>
          <publisher-loc>Cambridge</publisher-loc>
          <publisher-name>Cambridge University Press</publisher-name>
        </nlm-citation>
      </ref>
      <ref id="ref34">
        <label>34</label>
        <nlm-citation citation-type="book">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Clark</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Brennan</surname>
              <given-names>SE</given-names>
            </name>
          </person-group>
          <article-title>Grounding in communication</article-title>
          <source>Perspectives on Socially Shared Cognition</source>
          <year>1991</year>
          <publisher-loc>Washington, D.C., United States</publisher-loc>
          <publisher-name>American Psychological Association</publisher-name>
        </nlm-citation>
      </ref>
      <ref id="ref35">
        <label>35</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Scaife</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Rogers</surname>
              <given-names>Y</given-names>
            </name>
          </person-group>
          <article-title>External cognition: how do graphical representations work?</article-title>
          <source>Int J Hum Comput Stud</source>
          <year>1996</year>
          <month>8</month>
          <volume>45</volume>
          <issue>2</issue>
          <fpage>185</fpage>
          <lpage>213</lpage>
          <pub-id pub-id-type="doi">10.1006/ijhc.1996.0048</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref36">
        <label>36</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Zhang</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Norman</surname>
              <given-names>DA</given-names>
            </name>
          </person-group>
          <article-title>Representations in distributed cognitive tasks</article-title>
          <source>Cognit Sci Multidisciplin J</source>
          <year>1994</year>
          <volume>18</volume>
          <issue>1</issue>
          <fpage>87</fpage>
          <lpage>122</lpage>
          <pub-id pub-id-type="doi">10.1207/s15516709cog1801_3</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref37">
        <label>37</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Liu</surname>
              <given-names>LL</given-names>
            </name>
            <name name-style="western">
              <surname>Park</surname>
              <given-names>DC</given-names>
            </name>
          </person-group>
          <article-title>Aging and medical adherence: the use of automatic processes to achieve effortful things</article-title>
          <source>Psychol Aging</source>
          <year>2004</year>
          <month>06</month>
          <volume>19</volume>
          <issue>2</issue>
          <fpage>318</fpage>
          <lpage>25</lpage>
          <pub-id pub-id-type="doi">10.1037/0882-7974.19.2.318</pub-id>
          <pub-id pub-id-type="medline">15222825</pub-id>
          <pub-id pub-id-type="pii">2004-14948-008</pub-id>
        </nlm-citation>
      </ref>
    </ref-list>
  </back>
</article>
