Published on in Vol 7 (2023)

Preprints (earlier versions) of this paper are available at https://preprints.jmir.org/preprint/46659, first published .
Automated Diet Capture Using Voice Alerts and Speech Recognition on Smartphones: Pilot Usability and Acceptability Study

Automated Diet Capture Using Voice Alerts and Speech Recognition on Smartphones: Pilot Usability and Acceptability Study

Automated Diet Capture Using Voice Alerts and Speech Recognition on Smartphones: Pilot Usability and Acceptability Study

Original Paper

1Department of Electrical and Computer Engineering, Duke University, Durham, NC, United States

2Department of Computer Science and Engineering, Texas A & M University, College Station, TX, United States

3Department of Biomedical Engineering, Duke University, Durham, NC, United States

Corresponding Author:

Jessilyn Dunn, BSc, PhD

Department of Biomedical Engineering

Duke University

1427 FCIEMAS

Durham, NC, 27708

United States

Phone: 1 9196605131

Email: jessilyn.dunn@duke.edu


Background: Effective monitoring of dietary habits is critical for promoting healthy lifestyles and preventing or delaying the onset and progression of diet-related diseases, such as type 2 diabetes. Recent advances in speech recognition technologies and natural language processing present new possibilities for automated diet capture; however, further exploration is necessary to assess the usability and acceptability of such technologies for diet logging.

Objective: This study explores the usability and acceptability of speech recognition technologies and natural language processing for automated diet logging.

Methods: We designed and developed base2Diet—an iOS smartphone application that prompts users to log their food intake using voice or text. To compare the effectiveness of the 2 diet logging modes, we conducted a 28-day pilot study with 2 arms and 2 phases. A total of 18 participants were included in the study, with 9 participants in each arm (text: n=9, voice: n=9). During phase I of the study, all 18 participants received reminders for breakfast, lunch, and dinner at preselected times. At the beginning of phase II, all participants were given the option to choose 3 times during the day to receive 3 times daily reminders to log their food intake for the remainder of the phase, with the ability to modify the selected times at any point before the end of the study.

Results: The total number of distinct diet logging events per participant was 1.7 times higher in the voice arm than in the text arm (P=.03, unpaired t test). Similarly, the total number of active days per participant was 1.5 times higher in the voice arm than in the text arm (P=.04, unpaired t test). Furthermore, the text arm had a higher attrition rate than the voice arm, with only 1 participant dropping out of the study in the voice arm, while 5 participants dropped out in the text arm.

Conclusions: The results of this pilot study demonstrate the potential of voice technologies in automated diet capturing using smartphones. Our findings suggest that voice-based diet logging is more effective and better received by users compared to traditional text-based methods, underscoring the need for further research in this area. These insights carry significant implications for the development of more effective and accessible tools for monitoring dietary habits and promoting healthy lifestyle choices.

JMIR Form Res 2023;7:e46659

doi:10.2196/46659

Keywords



Diet-related diseases such as type 2 diabetes and coronary heart disease continue to increase at a staggering rate [1,2]. The International Diabetes Federation estimates that every 5 seconds, someone dies of diabetes or diabetes-related complications [3], and according to the World Health Organization, obesity prevalence around the world has nearly tripled since 1975 [4,5]. These global data highlight an urgency for innovative solutions to alleviate this problem. Incentivizing the adoption of a healthy diet can be instrumental in achieving positive outcomes [6-8]. Consequently, there have been many conversations concerning potential policy approaches such as taxing sugar-sweetened beverages [7,8] or reducing sodium levels [6,8] in processed foods to influence individuals’ adoption of a healthy diet. As for individualistic approaches, a proliferation of studies [9] demonstrates that diet monitoring carries substantial promise because it brings mindfulness [10] to eating. Mindfulness allows individuals to not only be aware of what they are eating but also identify when they are eating, food intolerances, positive and negative dietary habits, and potentially accelerated efforts toward a healthy lifestyle.

Gold-standard diet monitoring approaches include manual self-reporting methods such as 24-hour dietary recall, which are often highly inaccurate [9] and not user-friendly [9], leading to user attrition. In light of the shortcomings of conventional methods, there has been a growing interest in automatic dietary monitoring [9,11,12]. Automatic dietary monitoring systems leverage technology to monitor relevant aspects of food intake, such as timing and duration of meals, amount of food consumed, and nutritional content. In recent years, advances in mobile technologies have facilitated the proliferation of smartphone-assisted food-logging applications, such as text-based applications that rely on user-inputted text and computer vision-based applications that rely on barcode scans or photos taken by users to support food recognition and calorie estimation [13-15]. However, smartphone-assisted diet monitoring technologies suffer from low adherence [16] mainly because it is challenging for many individuals to remember to log their food intake [17]. Even with automated reminders and notifications, adherence remains low because the average user receives many other notifications from applications on their device [18], leading to alert fatigue. Furthermore, most of the alerts that individuals receive on their smartphones are silent or have generic reminder tones, which rarely attract the user’s attention. Diet logging is also often perceived as time-consuming [16,17], primarily because of the need for hands-on manual text or image and photo input, and the time required to retrieve each food item’s nutritional content. Accordingly, even with smartphone assistance, there is still a low long-term uptake of food-logging applications.

Accordingly, diet logging continues to be a challenge despite the existence of powerful technologies such as voice alerts, speech recognition (use of computers to detect human speech with the ultimate goal of translating it into text), and natural language processing (NLP—providing computers with an ability to parse and analyze human speech or text) that could change the landscape but have been less explored in this area [19-21]. This is especially important for improving diet recall and diet logging frequency, which could lead to more effective research and nutritional interventions for diet-related diseases such as type 2 diabetes.

This paper explores the usability and acceptability of speech recognition technologies for automated diet logging. We define automated diet logging as the act of acquiring a user’s spoken or typed natural language description of their consumed food, and using this to generate a time-stamped diet log comprising food item representations and automatically retrieved nutritional content (supported by NLP). We have developed base2Diet, an iOS application that employs voice-assisted technologies for hands-free diet logging. Some critical questions we strive to answer with base2Diet include the following: does voice-assisted diet logging improve diet logging adherence? Are users open to voice-assisted diet logging?


Application Development

We designed and developed base2Diet—an iOS smartphone application (Figures 1-3) for voice-assisted diet logging using Swift programming language on XCode—Apple’s integrated development environment. The app has 2 modes: text and voice. Text mode uses personalized text alerts to prompt users to log their food intake using text. In contrast, voice mode uses simultaneous text notifications and voice alerts to prompt users to log their food intake, which they can do using either text or voice. To send personalized alerts to users’ phones at designated times, we designed and developed a Node JS scheduling server hosted on Google Cloud. In addition, we used Firebase for user authentication and cloud storage. The app automatically creates time-stamped diet logs (Figure 2) based on user-provided natural language descriptions (voice or text) of the consumed food and the food’s nutritional content, which it obtains by querying Nutritionix [22]—an NLP-based application programming interface (API) that has been used in health applications such as the Vida Health app [23], and in several research studies [24,25]. Nutritionix’s NLP API allows users to describe their meals in everyday language and receive an accurate breakdown of the nutritional information of the consumed foods. For example, by simply stating, “I had 5 bacon slices, 1 plain bagel, a fried egg, a slice of cheddar cheese, and a glass of milk for breakfast,” the Nutritionix API will search for 5 items: bacon (5 slices), plain bagel (1), fried egg (1), cheddar cheese (1 slice), and glass of milk (1) and provide nutritional content for each of them based on the provided quantities. Conventional databases lacking NLP functionality would not understand that this query requires the retrieval of 5 distinct items from the database. To facilitate voice-assisted diet logging, we used Apple’s speech recognition framework [26] to detect and transcribe speech to text. Finally, we used the Evernote API [27] to connect base2Diet with Evernote, which we used for creating and storing time-stamped diet logs.

Figure 1. Screenshots of the base2Diet app. The leftmost screen shows the voice version of the app. It allows individuals to choose how they would like to log their food intake. The microphone button indicates choosing voice to log food intake, and the texting hand indicates choosing to text for diet logging. If an individual chooses voice, the middle view is presented. If the user chooses text, the rightmost view is presented.
Figure 2. Sample timestamped diet log automatically created by the base2Diet application in Evernote.
Figure 3. base2Diet app system diagram.

Study Design

We performed a 2-arm study with 18 participants (text: n=9, voice: n=9) and no crossover. We used a combination of self-selection and random sampling to allocate participants into groups. Strong preferences expressed by participants for a particular study arm were honored; otherwise, participants were randomly assigned to one of the study arms. Participants in the text arm only used text for diet logging, and they received personalized text alerts as reminders to log their food intake. Participants in the voice arm could use both voice and text to log their food intake, and they received personalized, simultaneous text notifications and voice alerts as reminders for them to log their food intake. Voice-assisted diet logging in the base2Diet app is both interactive and personalized. Once a user in the voice arm chooses to log their food intake using voice, the base2Diet app waits for the user to speak. If user’s speech is not detected within 8 seconds, the app says out loud, “Sorry X, I didn’t quite get that. Please try again,” where X is the user’s name. After this message, if no user’s speech is detected for another 4 seconds, the app closes the session and emits a failure notification sound. If speech is detected, base2Diet transcribes the speech to text, uses the transcribed speech text to query Nutritionix for the stated foods’ nutritional content, and automatically creates a time-stamped Evernote diet log comprising the transcribed speech text and the retrieved nutritional content (Figure 2). The base2Diet app displays the natural text transcribed from the user’s speech (Figure 1). If the user notices any discrepancies in the transcribed text, they can rerecord their food intake. On the text arm, once user text is detected, base2Diet immediately queries Nutritionix to retrieve the stated foods’ nutritional content before creating a time-stamped diet log in Evernote.

Personalized Alerts

To enhance user experience and increase engagement, we personalized all alerts. We used our custom server and cloud database with user information to personalize and send alerts. Each text notification on the base2Diet app addresses the user by name and refers to the appropriate time of day and the relevant meal for that time of day. For example, a breakfast alert for a participant named Jessica would present a text notification that says, “Good morning Jessica! What’s for breakfast?” To protect the user’s privacy, especially in public spaces, voice alerts did not mention the user’s name. Instead, they used prerecorded artificial intelligence–generated prompts [28] residing on the server. The user’s name would only be mentioned in the interactive dialogue between the user and the base2Diet app if the user is attempting to log their food intake using voice and the base2Diet app does not detect any speech. If no diet log is created 30 minutes after an alert is sent, only 1 additional reminder is sent to prevent intrusive app behavior. If food intake is logged before the set time for the reminder, base2Diet does not send any alerts. To support user privacy and freedom (the ability of a user to choose to exit unwanted actions or opt out of using any feature that is not of interest to the user at any given time, for example, voice alerts when a user is busy), base2Diet allows users to deactivate voice alerts for up to 1 full day. The system resets and reenables voice alerts every day at 4 AM. After each reset, users can manually deactivate the alerts for that day within the base2Diet app. When a user deactivates voice alerts, that user will receive text-only notifications but can use voice for food intake logging if they choose so.

Autopopulation of Nutritional Content

The base2Diet app leverages Nutritionix’s NLP API to convert the natural language description of the consumed food provided by the user in text or speech form into its nutritional information through food item retrieval from the Nutritionix database. Consequently, the base2Diet app logs the food and nutritional content into an Evernote-based food diary.

Data Collection Procedure

Flyers, social media, and word of mouth were used to recruit participants, who were assigned to either the text study arm (n=9) or the voice study arm (n=9). Though participants were located across the United States and were free to travel worldwide, the base2Diet app could detect their time zones and send alerts at the correct times. In the first phase of the study (phase I), all study participants received alerts at 9 AM, 12 PM, and 6 PM for the first 14 days. In the subsequent 14 days (phase II), participants were free to choose the times to receive the diet logging prompts, which they could change at any time during these 14 days.

On multiple occasions, we observed up to 5 distinct diet logs created by a user within a minute. Therefore, we defined distinct diet logging events as diet logging times separated by at least 30 minutes and otherwise considered multiple logs within a 30-minute span to be a part of a single-diet logging event. We defined an active day as a day when a participant creates at least 1 diet log by using the base2Diet app. Daily active users were defined as the number of users who log their food intake at least once per given study day. In phase II of the study, participants could choose when they wanted the app to remind them to log their food intake. We tracked how often users on both arms changed their alert times using Cloud Firestore [29]. We also enabled the participants to modify their diet logs at any time during the study and tracked this variable as well. Attrition rate was the rate at which participants dropped out of the study, defined as at least 7 consecutive days without the use of the base2Diet app. At the end of the study, all 18 participants completed a web-based usability survey that we developed, which had 14 multiple-choice questions, and 6 and 9 free-response questions in the text and voice arms, respectively. The survey was aimed at gauging participants’ perceptions of the usability and acceptability of voice technologies and NLP for automated diet capturing.

Statistical Analysis

The response variables included (1) the total number of distinct diet logging events per participant, and (2) the total number of active days per participant. The normality of these variables was assessed graphically as well as formally by using the Shapiro-Wilk normality test. A comparison between means of the voice and text arm was performed through unpaired t tests, with a significance threshold of P<.05. All statistical analyses were performed in R (version 4.0.2), and all graphs were generated in R (version 4.0.2) using ggplot2.

Ethics Approval

This study obtained ethical approval from the institutional review board at Duke University (protocol 2022-0236). Prior to enrollment, we obtained electronic informed consent from each study participant using REDCap.

All participants were provided a US $10 iTunes gift card as compensation for participating in the study. The disbursement of the gift cards was done at the end of the study after participants had filled out the exit survey.


Participant Characteristics

There were 18 participants in the study (8 males, 10 females; age range 20-39 years). A total of 16 participants were enrolled at Duke University as undergraduate or graduate students, 1 was a University of North Carolina student, and 1 was a working professional. Six of the participants were identified as Black, 8 identified as Asian, and 5 identified as White. The study inclusion criteria were being older than 18 years, currently living in the United States, being proficient in written and spoken English, owning an iPhone, and not maintaining a food diary at the start of the study.

User Engagement Metrics

To understand the differences between the impact of text versus voice alerts on diet logging behaviors, we used seven key performance indicators and user engagement metrics: (1) distinct diet logging events per participant, (2) total active days per participant, (3) daily active users, (4) feature usage, (5) attrition rate, (6) diet log modification, and (7) perceived usability. The number of distinct diet logging events per participant was higher in the voice arm than in the text arm (mean 30 vs 18, respectively; P=.03, unpaired t test; Figure 4). In addition, the total number of active days per participant was also higher in the voice arm than in the text arm (mean 19 vs 13, respectively; P=.04, unpaired t test; Figure 4). In general, the number of daily active users was higher in the voice arm as compared with the text arm (mean 6 vs 4), with the exception of 3 of the study days in which the number of active users in the voice arm was equal to that of the text arm (Figure 5).

The attrition rate in the text arm was higher than in the voice arm, with only 1 (11%) participant dropping out of the study in the voice arm and 5 (56%) in the text arm (Figures S1 and S2 in Multimedia Appendix 1).

We also explored the influence of preselected versus user-selected alert timing for diet logging reminders. Although the first 2 weeks of the study used preselected alert times based on typical US meal times of 9 AM, 12 PM, and 6 PM [30], in the second 2 weeks of the 4-week study, users selected their alert times, and had the option to change the alert times at any point over those 2 weeks. Of the 18 participants, 4 participants in the voice arm changed their alert times beyond the initial setup, whereas none of the text arm participants changed their alert times after the initial setup (Table 1).

All 18 study participants completed the usability survey that was disseminated at the end of the study. Our survey analysis showed that 67% (n=12) of the participants highly valued the app’s functionality of automatically populating the food items’ nutritional content while also allowing the participants to edit the created diet logs. Interestingly, however, no participant actually modified their diet logs during the study, even though all participants viewed the logs and had the option to change them. Additionally, 83% (n=15) of the participants either agreed or strongly agreed that hands-free speech-based diet logging makes it easier to adopt diet-monitoring habits (Figure S3 in Multimedia Appendix 1). When text arm participants were asked if they would have appreciated personalized voice alerts to remind them to log their meals, 56% (n=5) of them said yes (Figure S4 in Multimedia Appendix 1). When asked to elaborate, 1 participant said, “I find it harder to be on top of logging my diet with only text alerts.” When voice arm participants were asked about their perception of voice alerts, 78% (n=7) of them either agreed or strongly agreed that they liked voice alerts, but they also valued the ability to turn them off when they were busy (Figure S4 in Multimedia Appendix 1). When asked to elaborate, 1 participant said, “Voice alerts made me have regular meals, which was good. I now kinda look forward to the prompts so that I know it’s time for a meal.”

Figure 4. Box plot (A) shows the number of distinct diet logging events per participant by study arm (P=.03, unpaired t test). (B) shows the total active days per participant by study arm (P=.04, unpaired t test).
Figure 5. Total number of participants using the base2Diet app on each day of the study.
Table 1. Number of participants who changed their alert times per study arm.

Number of participants who changed alert times N times

0 times1 time5 times15 times
Voice Arm5211
Text Arm9000

Principal Findings

The primary objective of this study was to investigate the usability and acceptability of speech recognition technologies in automated diet logging. We designed and developed base2Diet—an iOS smartphone app for voice-assisted and automated diet logging. The results from the 28-day pilot study of base2Diet demonstrate that speech recognition technologies have tremendous potential to improve user experience and adherence to diet logging. Strikingly, the global market value for conversational artificial intelligence—chatbots or digital assistants that users can talk to—is expected to grow by 270%, from ≈US $6.8 million to ≈US $18.4 million between 2021 and 2026 [31]. The power of voice technologies to enhance user experiences and increase user engagement is among the key drivers of this growth. Our study, though a pilot in nature, supports the prevailing narrative, and our findings demonstrate that voice arm participants may be more likely to engage fully with the app than text arm participants (Table 1). Furthermore, a growing body of research points to the potential of voice technologies in improving diet recall and user experiences in diet monitoring [32-35]. For example, a study by Liang et al [35] revealed that 65% of younger and 60% of older participants preferred voice-assisted over web-based food recall. Beyond diet monitoring, voice technologies are being explored in numerous other domains [36-38], including assisted living [39].

Though the results of this study point to the promise of speech recognition technologies in automated diet logging, especially for improving user engagement and adherence, there is a need for user-centered design [40] to capture critical user needs, especially those pertaining to privacy and freedom. That 78% (n=7) of the participants in the voice arm either agreed or strongly agreed that they liked voice alerts, but also appreciated the ability to turn voice alerts off when busy (Figure S4 in Multimedia Appendix 1), reinforces previous findings that technologies should be adaptable to users’ preferences and the needs of the users should be central to the design process for technologies to be widely adopted. One participant’s experience clarifies this point. While elaborating why they appreciated speaking into their device to log their food intake, the participant said,

Dictating my meals to my phone was very convenient and took significantly less effort than typing. I love it as an interface, but I also loved having the option to type for [sic] when I was near others.

Follow-up conversations with participants revealed that all 18 participants viewed their diet logs at some time during the study. Although all participants had the option to manually edit their diet logs, none did it. We cannot be sure of why no one edited their diet logs; however, we do know that the average user inherently trusts autopopulated data or may not think it is worth the effort to change it, particularly when it comes to mobile apps [41,42]. Notably, the Instant Blood Pressure smartphone app [41], which measures blood pressure inaccurately, remained popular despite a disclaimer warning against its use for medical purposes. This highlights the significant influence that technology creators have over their users. As technology continues to become more ingrained in our daily lives, it is imperative for technology creators to acknowledge their ethical responsibility. Since the average user tends to trust information received from a mobile app without question, technology creators must ensure that the information and experiences they provide are accurate, nonharmful, and nonmisleading. Essentially, they must act ethically and prioritize the well-being of their users. Hence, it is crucial to raise awareness of sources of errors in autopopulated data to avoid errors being overlooked, particularly concerning app data used for health purposes.

Limitations

Apple’s speech recognition framework [26], which we used in the base2Diet app, is imperfect. Several participants, especially those with foreign accents, pointed out that the speech recognition framework was not always able to accurately transcribe their voices. This issue has also been brought to light by many other researchers [43,44], and there is a recognized need to improve the existing speech recognition frameworks to improve their translational impact. Continuously deploying applications that use these frameworks that perform poorly among minoritized populations may result in those groups of people rejecting such technologies due to an inherent assumption that they will not work for them. This has the potential to further exacerbate existing inequities in health and wellness.

Though Nutritionix has a comprehensive database containing 821,036 grocery items, 186,168 restaurant items, and 10,351 common food items, not every possible food item is represented [45]. Several researchers have highlighted the need for more comprehensive food databases [19,46] to improve outcomes and enhance user experiences within automated diet logging applications. This gap also points to the potential of machine learning methods capable of imputing nutritional content from the previously unseen food items [47].

In addition to the highlighted technology issues, this work has several other limitations. First, our study population was small and limited to relatively young adults. Other populations, such as working individuals, elder individuals, and individuals with type 2 diabetes, should be involved in future studies exploring voice-assisted diet logging technologies. In addition, individuals needed to own an iPhone to participate in the study, which might have inhibited participation by those from lower-income brackets since iPhones tend to be more expensive than mobile phones running other operating systems such as Android. Finally, while hands-free technologies facilitate convenient diet logging, primarily when users have limited or no hand function or are engaged in other activities, the base2Diet app could not provide a completely hands-free experience because Apple’s iOS operating system restricts the ability to unlock the phone using voice.

Future Research

Building on the promising results of this pilot study, there are several avenues for future research. One possible direction is to conduct large-scale, long-term trials with diverse populations, including individuals with specific dietary restrictions, or health conditions. This would provide a broader understanding of the generalizability and effectiveness of voice-assisted automated diet logging technologies in real-world settings.

Additionally, further investigations could be conducted to optimize the accuracy and efficiency of speech recognition technologies in capturing diet information, and to explore the integration of NLP techniques in food databases for a better understanding of user input. Further exploration of user preferences, needs, and privacy concerns in designing and developing these technologies could also provide valuable insights for optimizing user experience and adherence.

Furthermore, exploring the potential of voice alerts and just-in-time reminders as behavior change interventions could be an exciting area for future investigation. Overall, continued research in these areas has the potential to revolutionize the field of nutritional monitoring and interventions, improve public health outcomes, and address diet-related diseases such as type 2 diabetes.

Conclusions

This study paves the way for further research on integrating speech recognition and NLP in automated diet logging. We developed a proof-of-concept mobile app that demonstrates the possibilities of hands-free, in-the-moment, voice-assisted diet logging. Additionally, we explored interactive voice alerts and identified the need for respecting user privacy and freedom, further validating existing knowledge around the need to understand user preferences or needs and center the design process around the user.

Our findings indicate that voice-assisted automated diet logging technologies may lead to more effective nutritional monitoring and interventions for numerous diet-related diseases such as type 2 diabetes. Existing research has already established that diet monitoring can reverse the progression of type 2 diabetes or prevent or delay its onset in individuals with prediabetes [48-50]. Voice assistants and just-in-time reminders are additional technological layers that can improve the accuracy and effectiveness of diet-monitoring technologies and ultimately promote better health outcomes.

Acknowledgments

This study was funded by the National Science Foundation (NSF; subaward 1851173) for the Workshop on Technology for Automated Capture of Diet, Nutrition, and Eating Behaviors in Context. The funder played no role in the study design, data collection, analysis and interpretation of data, or the writing of this manuscript.

Data Availability

Deidentified data sets generated during this study, and deidentified survey responses are available from the corresponding author upon request. All correspondence and requests for study material should be directed to JD.

Authors' Contributions

JD, BM, and LC conceived of the idea and of the study design. LC developed the base2Diet mobile app and the scheduling server, performed analyses, generated visualizations, and drafted the manuscript. All authors read, edited, and approved the final version of the manuscript.

Conflicts of Interest

JD is a Scientific Advisor at Veri, Inc. BM is consultant for McAndrews Held & Malloy Ltd and a scientific consultant for Hugo Health for work unrelated to this manuscript. LC declares no financial competing interests.

Multimedia Appendix 1

Supplementary materials.

PDF File (Adobe PDF File), 1176 KB

  1. Roth GA, Mensah GA, Johnson CO, Addolorato G, Ammirati E, Baddour L, et al. Global burden of cardiovascular diseases and risk factors, 1990-2019: update from the GBD 2019 study. J Am Coll Cardiol 2020;76(25):2982-3021 [FREE Full text] [CrossRef] [Medline]
  2. Saeedi P, Petersohn I, Salpea P, Malanda B, Karuranga S, Unwin N, et al. Global and regional diabetes prevalence estimates for 2019 and projections for 2030 and 2045: results from the International Diabetes Federation Diabetes Atlas, 9th edition. Diabetes Res Clin Pract 2019;157:107843. [CrossRef] [Medline]
  3. IDF Diabetes Atlas, 10th edition.   URL: https://diabetesatlas.org/ [accessed 2022-10-23]
  4. Obesity and overweight. World Health Organization. 2021.   URL: https://www.who.int/news-room/fact-sheets/detail/obesity-and-overweight [accessed 2022-10-23]
  5. DeJesus RS, Croghan IT, Jacobson DJ, Fan C, St Sauver J. Incidence of obesity at 1 and 3 years among community dwelling adults: a population-based study. J Prim Care Community Health 2022;13:21501319211068632 [FREE Full text] [CrossRef] [Medline]
  6. Dodd R, Santos JA, Tan M, Campbell NRC, Mhurchu NC, Cobb L, et al. Effectiveness and feasibility of taxing salt and foods high in sodium: a systematic review of the evidence. Adv Nutr 2020;11(6):1616-1630 [FREE Full text] [CrossRef] [Medline]
  7. Roache SA, Gostin LO. The untapped power of soda taxes: incentivizing consumers, generating revenue, and altering corporate behavior. Int J Health Policy Manag 2017;6(9):489-493 [FREE Full text] [CrossRef] [Medline]
  8. Jacobson MF, Krieger J, Brownell KD. Potential policy approaches to address diet-related diseases. JAMA 2018;320(4):341-342. [CrossRef] [Medline]
  9. Prioleau T, Moore li E, Ghovanloo M. Unobtrusive and wearable systems for automatic dietary monitoring. IEEE Trans Biomed Eng 2017;64(9):2075-2089. [CrossRef] [Medline]
  10. Fung TT, Long MW, Hung P, Cheung LWY. An expanded model for mindful eating for health promotion and sustainability: issues and challenges for dietetics practice. J Acad Nutr Diet 2016;116(7):1081-1086. [CrossRef] [Medline]
  11. Hassannejad H, Matrella G, Ciampolini P, De Munari I, Mordonini M, Cagnoni S. Automatic diet monitoring: a review of computer vision and wearable sensor-based methods. Int J Food Sci Nutr 2017;68(6):656-670. [CrossRef] [Medline]
  12. Thomaz E, Essa IA, Abowd GD. Challenges opportunities in automated detection of eating activity. In: Rehg JM, Murphy SA, Kumar S, editors. Mobile Health: Sensors, Analytic Methods, and Applications. Cham: Springer; 2017:151-174.
  13. MyPlate calorie counter. MyPlate.   URL: https://www.myplateapp.com/ [accessed 2023-01-20]
  14. MyFitnessPal.   URL: https://www.myfitnesspal.com [accessed 2023-01-20]
  15. Bitesnap - Photo Food Journal.   URL: https://getbitesnap.com [accessed 2023-01-20]
  16. Ghelani DP, Moran LJ, Johnson C, Mousa A, Naderpoor N. Mobile apps for weight management: a review of the latest evidence to inform practice. Front Endocrinol (Lausanne) 2020;11:412 [FREE Full text] [CrossRef] [Medline]
  17. Cordeiro F, Bales E, Cherry E, Fogarty J. Rethinking the mobile food journal: exploring opportunities for lightweight photo-based capture. 2015 Presented at: CHI '15: Proceedings of the 33rd Annual ACM Conference on Human Factors in Computing Systems; April 18-23, 2015; Seoul Republic of Korea p. 3207-3216. [CrossRef]
  18. da Silva AVD, Vieira V. Towards an API for user attention prediction in mobile notification overload. In: Anais Estendidos Do XXIV Simpósio Brasileiro de Sistemas Multimídia e Web. Porto Alegre: Sociedade Brasileira de Computação - SBC; 2018:13-17.
  19. Fuchs KL, Haldimann M, Vuckovac D, Ilic A. Automation of data collection techniques for recording food intake: a review of publicly available and well-adopted diet apps. 2018 Presented at: 2018 International Conference on Information and Communication Technology Convergence (ICTC); October 17-19, 2018; Jeju, Korea (South) p. 58-65. [CrossRef]
  20. Naphtal R. Natural language processing based nutritional application (Thesis). Massachusetts: Department of Electrical Engineering and Computer Science, Massachusetts Institute of Technology; 2015.
  21. He Y, Hakguder Z, Shi X. Smart diet management through food image cooking recipe analysis. In: Mazzeo PL, Frontoni E, Sclaroff S, Distante C, editors. Image Analysis and Processing. ICIAP 2022 Workshops. Lecture Notes in Computer Science, Vol 13373. Cham: Springer; 2022:82-93.
  22. Nutrition API by Nutritionix.   URL: https://tinyurl.com/3dbemc7j [accessed 2022-12-17]
  23. Vida Health - Home.   URL: https://www.vida.com/ [accessed 2023-01-22]
  24. Miller HN, Berger MB, Askew S, Kay MC, Hopkins CM, Iragavarapu MS, et al. The nourish protocol: a digital health randomized controlled trial to promote the DASH eating pattern among adults with hypertension. Contemp Clin Trials 2021;109:106539 [FREE Full text] [CrossRef] [Medline]
  25. Li J, Guerrero R, Pavlovic V. Deep cooking: predicting relative food ingredient amounts from images. 2019 Presented at: Proceedings of the 5th International Workshop on Multimedia Assisted Dietary Management; October 21, 2019; Nice France p. 2-6. [CrossRef]
  26. Speech. Apple Developer Documentation.   URL: https://developer.apple.com/documentation/speech [accessed 2022-12-29]
  27. Documentation - evernote developers. Evernote.   URL: https://dev.evernote.com/doc/ [accessed 2022-12-17]
  28. Voicemaker® - Text to speech converter.   URL: https://voicemaker.in/ [accessed 2023-02-07]
  29. Firestore. Firebase.   URL: https://firebase.google.com/docs/firestore [accessed 2023-01-09]
  30. 40-Year trends in meal and snack eating behaviors of american adults. Elsevier Enhanced Reader.   URL: https://tinyurl.com/2ftm7222 [accessed 2023-01-11]
  31. Conversational AI market: global forecast to 2026. MarketsandMarkets Knowledge Store.   URL: https://tinyurl.com/ms3atbke [accessed 2022-11-21]
  32. Millard LAC, Johnson L, Neaves SR, Flach PA, Tilling K, Lawlor DA. Collecting food and drink intake data with voice input: development, usability, and acceptability study. JMIR Mhealth Uhealth 2023;11:e41117 [FREE Full text] [CrossRef] [Medline]
  33. Hezarjaribi N, Mazrouee S, Ghasemzadeh H. Speech2Health: a mobile framework for monitoring dietary composition from spoken data. IEEE J Biomed Health Inform 2018;22(1):252-264. [CrossRef] [Medline]
  34. Liu Y, Chen C, Lin Y, Chen H, Irianti D, Jen T, et al. Design and usability evaluation of mobile voice-added food reporting for elderly people: randomized controlled trial. JMIR Mhealth Uhealth 2020;8(9):e20317 [FREE Full text] [CrossRef] [Medline]
  35. Liang X, Batsis J, Yuan J, Zhu Y, Driesse T, Schultz J. Voice-assisted food recall using voice assistants. In: Duffy VG, Gao Q, Zhou J, Antona M, Stephanidis C, editors. HCI International 2022 - HCI for Health, Well-being, Universal Access and Healthy Aging. Lecture Notes in Computer Science, vol 13521. Cham: Springer; 2022:92-107.
  36. Shreya P, Shreyas N, Pushya D, Reddy N UM. BLIND ASSIST: a one stop mobile application for the visually impaired. 2021 Presented at: 2021 IEEE Pune Section International Conference (PuneCon); December 16-19, 2021; Pune, India. [CrossRef]
  37. Byonanebye DM, Nabaggala MS, Naggirinya AB, Lamorde M, Oseku E, King R, et al. An interactive voice response software to improve the quality of life of people living with HIV in Uganda: randomized controlled trial. JMIR Mhealth Uhealth 2021;9(2):e22229 [FREE Full text] [CrossRef] [Medline]
  38. Brinkel J, May J, Krumkamp R, Lamshöft M, Kreuels B, Owusu-Dabo E, et al. Mobile phone-based interactive voice response as a tool for improving access to healthcare in remote areas in Ghana - an evaluation of user experiences. Trop Med Int Health 2017;22(5):622-630. [CrossRef] [Medline]
  39. Pradhan A, Lazar A, Findlater L. Use of intelligent voice assistants by older adults with low technology use. ACM Trans Comput-Hum. Interact 2020;27(4):1-27. [CrossRef]
  40. Lowdermilk T. User-Centered Design: A Developer's Guide to Building User-Friendly Applications. Sebastopol, California: O'Reilly Media, Inc; 2013.
  41. Plante TB, O'Kelly AC, Macfarlane ZT, Urrea B, Appel LJ, Miller Iii ER, et al. Trends in user ratings and reviews of a popular yet inaccurate blood pressure-measuring smartphone app. J Am Med Inform Assoc 2018;25(8):1074-1079 [FREE Full text] [CrossRef] [Medline]
  42. Hohmann-Marriott B, Starling L. “What if it's wrong?” Ovulation and fertility understanding of menstrual app users. SSM - Qualitative Research in Health 2022;2:100057. [CrossRef]
  43. Lima L, Furtado V, Furtado E, Almeida V. Empirical analysis of bias in voice-based personal assistants. 2019 Presented at: Companion Proceedings of The 2019 World Wide Web Conference; May 13-17, 2019; San Francisco, USA p. 533-538. [CrossRef]
  44. Palanica A, Thommandram A, Lee A, Li M, Fossat Y. Do you understand the words that are comin outta my mouth? Voice assistant comprehension of medication names. npj Digit Med 2019;2(1):55. [CrossRef]
  45. Nutritionix - largest verified nutrition database.   URL: https://www.nutritionix.com/ [accessed 2022-12-17]
  46. Lim J, Lim C, Ibrahim I, Syahrul J, Mohamed Zabil MH, Zakaria NF, et al. Limitations of existing dialysis diet apps in promoting user engagement and patient self-management: quantitative content analysis study. JMIR Mhealth Uhealth 2020;8(6):e13808 [FREE Full text] [CrossRef] [Medline]
  47. D'Arcy J, Qi S, Steinberg D, Dunn J. Semantic nutrition: estimating nutrition with mobile assistants. Machine Learning for Healthcare. 2020.   URL: https:/​/static1.​squarespace.com/​static/​59d5ac1780bd5ef9c396eda6/​t/​5f244db1095aca17497b7244/​1596214706014/​29_CameraReadySubmission_Abstract_Final.​pdf [accessed 2023-02-16]
  48. Salas-Salvadó J, Martinez-González MÁ, Bulló M, Ros E. The role of diet in the prevention of type 2 diabetes. Nutr Metab Cardiovasc Dis 2011;21 Suppl 2:B32-B48. [CrossRef] [Medline]
  49. Hall K, Chung S. Low-carbohydrate diets for the treatment of obesity and type 2 diabetes. Curr Opin Clin Nutr Metab Care 2018;21(4):308-312. [CrossRef] [Medline]
  50. Hallberg SJ, Gershuni VM, Hazbun TL, Athinarayanan SJ. Reversing type 2 diabetes: a narrative review of the evidence. Nutrients 2019;11(4):766 [FREE Full text] [CrossRef] [Medline]


API: application programming interface
NLP: natural language processing


Edited by A Mavragani; submitted 20.02.23; peer-reviewed by S Rego, D Chrimes; comments to author 03.04.23; revised version received 10.04.23; accepted 11.04.23; published 16.05.23

Copyright

©Lucy Chikwetu, Shaundra Daily, Bobak J Mortazavi, Jessilyn Dunn. Originally published in JMIR Formative Research (https://formative.jmir.org), 16.05.2023.

This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR Formative Research, is properly cited. The complete bibliographic information, a link to the original publication on https://formative.jmir.org, as well as this copyright and license information must be included.