Abstract
Objective:
A significant gap exists between those who need and those who receive care for eating disorders (EDs). Novel solutions are needed to encourage service use and address treatment barriers. This study developed and evaluated the usability of a chatbot designed for pairing with online ED screening. The tool aimed to promote mental health service utilization by improving motivation for treatment and self-efficacy among individuals with EDs.
Method:
A chatbot prototype, Alex, was designed using decision trees and theoretically-informed components: psychoeducation, motivational interviewing, personalized recommendations, and repeated administration. Usability testing was conducted over four iterative cycles, with user feedback informing refinements to the next iteration. Post-testing, participants (N=21) completed the System Usability Scale (SUS), the Usefulness, Satisfaction, and Ease of Use Questionnaire (USE), and a semi-structured interview.
Results:
Interview feedback detailed chatbot aspects participants enjoyed and aspects necessitating improvement. Feedback converged on four themes: user experience, chatbot qualities, chatbot content, and ease of use. Following refinements, users described Alex as humanlike, supportive, and encouraging. Content was perceived as novel and personally relevant. USE scores across domains were generally above average (~5 out of 7), and SUS scores indicated “good” to “excellent” usability across cycles, with the final iteration receiving the highest average score.
Discussion:
Overall, participants generally reflected positively on interactions with Alex, including the initial version. Refinements between cycles further improved user experiences. This study provides preliminary evidence of the feasibility and acceptance of a chatbot designed to promote motivation for and use of services among individuals with EDs.
Keywords: eating disorder, chatbot, mental health treatment, digital intervention, health screening, conversational agent, chatbot design, chatbot development, mhealth
Introduction
Eating disorders (EDs) are complex mental illnesses characterized by debilitating, pervasive psychological and physiological consequences when left untreated. Given that lack of or delay in treatment initiation results in poorer prognosis and higher relapse risk (American Psychiatric Association, 2006), novel solutions are needed to facilitate treatment engagement. Such solutions must address the most prominent barriers patients face in seeking and receiving prompt treatment, including stereotypes surrounding EDs, denial of illness severity, compromised motivation for treatment, and knowledge deficits regarding treatment resources (Ali et al., 2017; Ali et al., 2020).
We partnered with the National Eating Disorders Association (NEDA) in the U.S. to disseminate an evidence-based online EDs screen, which garners about 200,000 respondents annually. Our prior work showed that while most respondents (86%) screened positive for an ED, few had previously received treatment (14%) or were currently receiving treatment (3%) (Fitzsimmons-Craft et al., 2019), which is consistent with other work suggesting less than 20% of those with EDs ever receive treatment (Eisenberg, Nicklett, Roeder, & Kirz, 2011; Kazdin, Fitzsimmons-Craft, & Wilfley, 2017). Additional investigation suggested that only 16% of NEDA screen completers who screened positive for an ED and were not in treatment had initiated care following screen completion, with this figure likely an overestimate given its basis on only those who provided follow-up data (Fitzsimmons-Craft et al., 2020). Thus, additional tools are needed to increase motivation for ED treatment and service utilization.
Digital tools offer a user-friendly, cost-effective, and accessible solution to tackle patient concerns and to broadly disseminate mental health-related interventions. Such tools include chatbots, which are computer programs that simulate human conversations (Torous et al. 2021). Despite having awareness of conversating with a robot, humans still respond to chatbots as they would to humans, as evidence supports (Gratzer et al., 2021; Montenegro, da Costa, & da Rosa Righi, 2019), which reflects the success chatbots have shown in establishing digital therapeutic alliances, relationships which may mediate the effectiveness of digital mental health tools by increasing user trust, fidelity, and engagement (Torous & Haim, 2018). In addition, chatbots may improve mental health symptoms like body image (Car et al., 2020; Fitzsimmons-Craft et al., 2022) and promote positive health-related behavior change (Car et al., 2020; Pereira & Díaz, 2019; Zhang, Oh, Lange, Yu, & Fukuoka, 2020).
These findings and others suggest that a chatbot may be appropriate to facilitate motivation for and utilization of services in individuals with EDs. As such, this study developed and evaluated the usability of a chatbot designed for that purpose and to be used following online screening. We detail the process of developing this chatbot using user-centered design and multiple iterative cycles of usability testing (Graham et al., 2019; Lyon & Bruns, 2019).
Method
Participants and Recruitment
Participants were recruited via digital ads on social media platforms, physical flyers posted throughout St. Louis, Missouri, and an email listserv of individuals interested in research. Advertisements directed participants to Qualtrics, which administered an online eligibility questionnaire and collected demographic information. Eligibility criteria required that participants were at least 18 years old, had screened positive for a clinical or subclinical ED based on the Stanford-Washington University ED Screen (SWED) (Graham, Trockel et al., 2019), reported not currently being in treatment for an ED, and owned a mobile phone. Eligible participants were asked to provide contact information and were subsequently invited to participate by a member of our research team.
Chatbot Design
The chatbot, named Alex, was developed using a user-centered design approach, which includes an iterative, six phase process (investigate, ideate, prototype, evaluate, refine, develop) aimed to produce technology that integrates feedback from stakeholders and members of the target population throughout development (Graham, Wildes et al., 2019). This study is part one of a three-study series following the Multiphase Optimization Strategy (MOST) framework, which ensures that interventions are optimized before they are tested in an RCT (Collins, 2018), and which includes three phases: 1) preparation; 2) optimization; and 3) evaluation. The final version of Alex from this study will be tested in study two (the optimization phase) to assess the effectiveness of chatbot components on behavior change and help-seeking (NCT04806165). Study three will follow, which will include an RCT using the optimized bot. All procedures were approved by the institutional review board.
Alex was hosted by a private mental health chatbot company, X2AI, and featured a fully automated, synchronous conversation delivered via SMS or Facebook Messenger. Alex was developed using decision trees, as part of a rule-based approach that is more time- and cost-efficient, revisable, and controllable than an artificial-intelligence approach (Boucher et al., 2021). Decision trees use “if-then” statements to navigate conversations via pre-programmed logic, including conditional statements and branching patterns (Figure 1), and recognize specific keywords, simple responses, and common phrases (e.g., “yes” or “I don’t know”) in users’ messages (Buntine, 1992). Given that users in past rule-based chatbot studies commonly perceived the tools as “robotic” (Abd-Alrazaq et al., 2021; Denecke, Abd-Alrazaq, & Househ, 2021), several guiding principles informed Alex’s design, each intended to provide users with a human-like, amiable experience: 1) chatbot responses were kept short in length and included colloquial, empathetic, and congenial rhetoric, and 2) chatbot responses were limited to four consecutive messages.
Fig. 1.
Screens illustrate user conversations with Alex for each component tested. From left to right: psychoeducation, motivational interviewing, personalized recommendations, and repeated administration.
Chatbot Components
Alex started with a brief introduction covering what to expect, protocol for crisis, and limitations of the tool. Four theoretically-informed components (i.e., psychoeducation (PE), motivational interviewing (MI), personalized recommendations for services (PR), and repeated administration (RA)) followed (Figure 2).
Fig. 2.
Four components were tested: psychoeducation, personalized recommendations for intervention, motivational interviewing, and repeated administration. Each component targets theoretically-informed mechanisms for increasing motivation for treatment and self-efficacy. Modules ultimately aimed to improve uptake of mental health services.
The PE component was based on the mental health literacy model, which has been shown to improve patients’ help-seeking attitudes and treatment need perceptions (Xu et al., 2018). Following the Academy for Eating Disorders’ “Nine Truths about EDs” (Schaumberg et al., 2017) this component refuted common ED stereotypes and myths, informed users of ED health consequences, provided tailored information for each specific ED psychopathology (e.g., restriction, compensatory behaviors, etc.), and emphasized the seriousness of user concerns.
The MI component maps closely onto cognitive dissonance theory (Festinger, 1962) and self-perception theories (Bem, 1967), and was guided by Miller and Rollnick’s (2012) core elements of MI. MI highlights discrepancies between individuals’ unhealthy behaviors and their healthy goals (Lundahl & Burke, 2009), which has been shown to improve outcomes and uptake of care by improving self-efficacy and motivation to change (Copeland, McNamara, Kelson, & Simpson, 2015). A review of meta-analytic findings showed that MI is significantly more effective than no treatment (ds up to .57) for a range of problems and for increasing treatment engagement (Burke, Arkowitz, & Menchola, 2003; Lundahl & Burke, 2009).
The PR component provided personalized recommendations for seeking intervention. The module was based on the Elaboration Likelihood Model (ELM) (Petty & Cacioppo, 1981), whereby individuals are more likely to process information in an active and thoughtful manner when they find messages personally relevant. Messages that are processed in this way are retained longer and tend to improve individuals’ motivation and self-efficacy (Kreuter & Wray, 2003). In this component, participants responded to questions regarding treatment preferences (in-person or online) (Figure 3), and Alex followed-up with tailored resources recommended from the NEDA website (in contrast to a lengthy list of options often presented at the end of online screens). Findings have indicated that tailored messages have a greater impact on health behavior versus a control group (r=.074) (Noar, Benac, & Harris, 2007), and that text message-based health promotion messages have an effect of d=.33, with greater efficacy for those using personalization (p=.001) (Head, Noar, Iannarino, & Harrington, 2013).
Fig. 3.
The personalized recommendations component offered users with tailored resource and service recommendations based on items assessing treatment preferences. This figure outlines the branching logic used by Alex to determine which resources and services to provide users.
* =Iterative refinements added pathways for users to see both in-person/telehealth and online options, and multiple options from each
**=Additional online resource options were added between cycles, including references to online support groups and an online peer-to-peer support community
The RA component provided up to three interactive check-ins which reminded users of resources and services available to them and promoted reflection on overcoming barriers. Three days after completing the main conversation, users received their first check-in. Those who endorsed seeking help stopped receiving RA check ins, while those users who did not continued to receive up to two more check-ins each spaced three days apart until they reported services seeking or utilization. Reminders have been shown to increase adherence to health-related interventions and treatment (Fenerty, West, Davis, Kaplan, & Feldman, 2012). For example, meta-analytic findings have indicated that reminder systems improve medication adherence in chronic disease (d=.41) (Thakkar et al., 2016) and other issues (Misono et al., 2010).
The introduction, PE, MI, and PR components were designed for completion in 15-20 minutes. RA check-ins took ~3 minutes to complete.
Procedure
In-person and remote usability testing was conducted with four sequential testing groups, as deemed optimal when developing mHealth interventions using user-centered design (Molina-Recio et al., 2020). The first two cycles were conducted in-person and were moderated by a research team member, who supervised the participants during testing to catch technical glitches, provide instruction when needed, and capture detailed feedback (Wozney et al., 2016). This ensured that necessary revisions were made in order for users in consequent remote, unmoderated cycles to be well-equipped to independently navigate the tool.
Cycles 1 & 2.
Eligible participants near St. Louis, Missouri were invited to our lab where informed consent was obtained and testing sessions were conducted with a team member. Via SMS or Facebook Messenger, participants first assessed the PE, MI, and PR components of the chatbot using the think-aloud protocol, which provided instantaneous reactions, thoughts, and opinions of participants during testing (Jaspers, Steen, van den Bos, & Geenen, 2004). The RA component was not tested in this cycle. A 60-minute semi-structured interview immediately followed. Participants were audio-recorded throughout the testing session and interview. Post-interview, participants completed an online questionnaire, including the measures described below, as well as the Eating Disorder Examination-Questionnaire (EDE-Q) (Fairburn & Beglin, 1994) to assess ED psychopathology. Participants were compensated with a $20 Amazon electronic gift card.
Cycles 3 & 4.
Remote, unsupervised usability testing in Cycles 3 and 4 was conducted over two weeks. Eligible participants throughout the United States were invited to participate via a Qualtrics survey link, which hosted an informed consent form and a baseline survey collecting demographic information and EDE-Q results. After completing these items, users initiated the SMS conversation by entering the unique start code IDs assigned by Qualtrics’ random number generator. These cycles completed each component tested by Cycles 1 and 2, in addition to the RA component. After two weeks, participants scheduled a ~60 minute semi-structured interview call and completed questionnaires, which included the quantitative measures below. Interviews were conducted by a research team member and were audio recorded. Participants were compensated with a $30 Amazon electronic gift card.
Measures
Quantitative Data
The System Usability Scale (SUS) was administered post-engagement to assess usability (Brooke, 1995). The scale consists of 10-items scored on a 5-point Likert scale, with response options ranging from strongly disagree (1) to strongly agree (5). SUS scores range from 0-100, with an evaluation of 500 studies delineating a score over 68 illustrating “above average” usability and demonstrable user acceptance (Sauro, 2011). The scale is validated for use in small sample sizes (Bangor, Kortum, & Miller, 2008; Lewis & Sauro, 2009).
Users also completed The Usefulness, Satisfaction, and Ease of Use (USE) Questionnaire Short-Form, which measures four dimensions: usefulness, ease of use, ease of learning, and satisfaction of users (Lund, 2001). Items are scored along a 7-point Likert scale, with response options ranging from strongly disagree (1) to strongly agree (7). Subscale scores are generated by averaging item scores independently on each dimension.
Qualitative Feedback
Semi-structured interview questions gathered user perceptions of each chatbot component (PE, MI, PR, RA), design of the chatbot, usefulness, and overall positive and negative experiences. See the full interview guide in the Online Supplement.
Analytic Strategy
Descriptive Analysis
Descriptive statistics were calculated for each cycle using the Statistical Package for Social Sciences (SPSS) version 17.
Qualitative Analysis
Iterative Development.
The initial version of Alex was tested by Cycle 1 participants. Interviews were transcribed and qualitative feedback was assessed by frequency and feasibility to inform refinements to Alex between testing cycles. Some suggestions, including improving the chatbot’s recognition of complex language and enabling users to schedule reminders in the RA component, were not feasible given Alex’s rule-based design.
Thematic Analysis.
Qualitative data was classified following the six-phase thematic analysis framework outlined by Braun and Clarke (2006), which strikes balance between flexibility and rigor, and provides opportunity to evaluate latent themes. Audio recordings were transcribed independently by one team member (J.S.) and later checked independently by team member (B.D.). Interview responses were categorized by component or as overall experience feedback. Using methods by Hannah and Lautsch (2011), J.S. and B.D. privately generated preliminary codes and met at multiple points to compare, evaluate, and discuss similarities and differences. Once members reached a consensus, codes were categorized as positive or negative feedback and then sorted into themes. The final codebook presents positive and negative user feedback, including changes in user sentiment toward the chatbot following iterative refinements.
Results
Participants
Qualitative measures require a sample size above nine to achieve coding saturation and between 16-24 to achieve meaning saturation (Hennink, Kaiser, & Marconi, 2017; Creswell & Creswell, 2017). Therefore, a total of 21 participants were enrolled (n=5, Cycles 1-3; n=6, Cycle 4) (Table 1), and no additional participants were considered. Participants largely identified as white (non-Hispanic) (n = 17, 81%) and female (n=20, 95%), with one participant of male sex included in Cycle 4. Represented ED diagnoses based on SWED criteria included unspecified feeding or eating disorder (UFED) (n=8, 38%), clinical or subclinical bulimia nervosa (BN) (n = 6, 29%), clinical or subclinical binge eating disorder (BED) (n = 5, 24%), anorexia nervosa (AN) (n = 1, 5%), and purging disorder (PD) (n = 1, 5%). The mean EDE-Q global score amongst participants (4.48, SD=1.29) was above the clinical cutoff (4) (Lund, 2001), confirming high ED psychopathology among users. See Table 1 for full details on study participants.
Table 1.
Table of Participant Characteristics by Usability Testing Cycle
Participant # | Race and Ethnicity | Sex | SWED-Diagnosis | EDE-Q Global Score (Mean ± SD) |
|
---|---|---|---|---|---|
Cycle 1 (n=5) | 3.59±1.02 | ||||
P1 | White | Female | UFED | ||
P2 | White, Hispanic | Female | UFED | ||
P3 | White | Female | UFED | ||
P4 | White | Female | UFED | ||
P5 | Asian | Female | UFED | ||
Cycle 2 (n=5) | 3.94±1.03 | ||||
P6 | Asian, White | Female | UFED | ||
P7 | White | Female | UFED | ||
P8 | White | Female | Subclinical Binge Eating Disorder | ||
P9 | White | Female | Subclinical Bulimia Nervosa | ||
P10 | Asian | Female | Subclinical Binge Eating Disorder | ||
Cycle 3 (n=5) | 4.88±0.42 | ||||
P11 | Hispanic | Female | Anorexia Nervosa | ||
P12 | White, Hispanic | Female | Purging Disorder | ||
P13 | White | Female | Binge Eating Disorder | ||
P14 | Black or African American, White | Female | Subclinical Bulimia Nervosa | ||
P15 | White | Female | Bulimia Nervosa | ||
Cycle 4 (n=6) | 3.51±1.47 | ||||
P16 | Black or African American, White | Female | Bulimia Nervosa | ||
P17 | White | Female | Bulimia Nervosa | ||
P18 | White | Female | Subclinical Binge Eating Disorder | ||
P19 | White | Female | Bulimia Nervosa | ||
P20 | White | Male | UFED | ||
P21 | Black or African American | Female | Subclinical Binge Eating Disorder |
|
|
Total (n=21) | 4.48 ±1.29 |
Note. UFED = unspecified feeding or eating disorder, EDE-Q= Eating Disorder Examination-Questionnaire, SWED= Stanford-Washington University ED Screen
Scale Scores
The average SUS score from Cycle 1 (83.0) indicated the initial intervention had above average usability and demonstrable user acceptance even prior to any refinements (Table 2). Average SUS scores across cycles ranged from 75.0-85.8. Though there was not a sizable change in average SUS scores between the first (83.0) and final (85.8) iteration, scores over all cycles were comparable or higher than scores obtained in other mental health chatbot usability studies (Cameron et al., 2019, Islam et al., 2021, Oh et al., 2021). Overall, Alex was rated positively based on USE scores, which were above 4 (out of 7) on most domains. High USE scores for the first iteration may explain why refinements did not drastically improve USE scores (Table 2).
Table 2.
Descriptive Statistics for SUS and USE Questionnaires by Usability Testing Cycle
Cycle | Descriptive | ||||
---|---|---|---|---|---|
SUS Total | Mean | SD | Min. | Max. | |
1 | 83.0 | 13.2 | 62.5 | 97.5 | |
2 | 77.0 | 16.7 | 50.0 | 90.0 | |
3 | 75.0 | 19.8 | 50.0 | 95.0 | |
4 | 85.8 | 10.2 | 70.0 | 97.5 | |
USE Usefulness | |||||
1 | 5.13 | 0.95 | 3.83 | 6.33 | |
2 | 4.67 | 0.89 | 3.33 | 5.67 | |
3 | 4.07 | 1.60 | 1.50 | 5.50 | |
4 | 5.00 | 1.06 | 3.33 | 6.17 | |
USE Ease of Use | |||||
1 | 6.12 | 0.66 | 5.40 | 7.00 | |
2 | 5.84 | 0.33 | 5.40 | 6.20 | |
3 | 4.68 | 2.09 | 1.60 | 6.80 | |
4 | 5.73 | 1.26 | 3.40 | 7.00 | |
USE Ease of Learning | |||||
1 | 6.33 | 0.62 | 5.67 | 7.00 | |
2 | 5.80 | 0.96 | 4.67 | 7.00 | |
3 | 5.53 | 2.06 | 2.00 | 7.00 | |
4 | 6.39 | 0.57 | 5.67 | 7.00 | |
USE Satisfaction | |||||
1 | 5.32 | 1.27 | 3.40 | 6.80 | |
2 | 4.52 | 1.60 | 1.80 | 5.80 | |
3 | 3.84 | 1.35 | 1.60 | 5.20 | |
4 | 4.80 | 1.46 | 3.00 | 6.60 |
Note. SUS= System Usability Scale (SUS), USE=Usefulness, Satisfaction, and Ease of Use Questionnaire
Thematic Analysis
Feedback from participants was labeled as “positive” or “negative”, with the latter identifying suggestions to improve the chatbot and which informed modifications between iterations (Table 4). Initial codes converged under the following general themes: user experience (subjective user perception and feelings), chatbot qualities (traits and characteristics of the bot as perceived by users, such as human-like or automated), chatbot content (overall and module-specific user reactions to information and resources), and ease of use (navigation and technical components) (Table 3).
Table 3.
Themes, Descriptions, and Feedback from Users across Usability Testing Cycles
Themes | Description: | Illustrative Quotes (P#) |
---|---|---|
Overall Experience | ||
Positive | ||
Experience |
|
|
Chatbot Qualities |
|
|
Content | N/A | N/A |
Ease of Use |
|
|
Negative | ||
Experience |
|
|
Chatbot Qualities |
|
|
Content | N/A | N/A |
Ease of Use | N/A | N/A |
Psychoeducation | ||
Positive | ||
Experience |
|
|
Chatbot Qualities | N/A | N/A |
Content |
|
|
Ease of Use | N/A | N/A |
Negative | ||
Experience |
|
|
Chatbot Qualities |
|
|
Content |
|
|
Ease of Use | N/A | N/A |
Motivational Interviewing | ||
Positive | ||
Experience |
|
|
Chatbot Qualities |
|
|
Content |
|
|
Ease of Use | N/A | N/A |
Negative | ||
Experience |
|
|
Chatbot Qualities | N/A | N/A |
Content |
|
|
Ease of Use | N/A | N/A |
Personalized Recommendations | ||
Positive | ||
Experience |
|
|
Chatbot Qualities | N/A | N/A |
Content |
|
|
Ease of Use | N/A | N/A |
Negative | ||
Experience |
|
|
Chatbot Qualities | N/A | N/A |
Content |
|
|
Ease of Use | N/A | N/A |
Repeated Administration | ||
Positive | ||
Experience |
|
|
Chatbot Qualities | N/A | |
Content |
|
|
Ease of Use | N/A | |
Negative | ||
Experience | N/A | |
Chatbot Qualities |
|
|
Content | N/A | |
Ease of Use | N/A |
Overall Experience
Users largely reported having enjoyable, gratifying experiences across cycles. Most users affirmed that chatbot content reinforced help-seeking, personal reflection, and behavior change. Participants did not report issues navigating the tool, reporting that it required little-to-no instruction: “…If you can chat with someone, you can do this” (P16). Of significance is how human-like most users described Alex as, using terms like “engaging”, “impartial”, and “realistic”. These qualities were attributed to Alex’s empathetic and comforting language: a user (P19) reported, “It also wasn’t super clinical, and it used emojis and things like that to make the experience more comfortable and to feel more normal.” Another user (P7) reported, “There were affirmations, and it backs up and restates what you tell her, so it was very spot-on. It was able to follow and help make connections.” Some users even found that removing human interaction compelled them to respond more honestly than if they were seeking in-person help: “I like that you can be more honest with your feelings because you're not face to face with somebody” (P4).
Some users in early cycles did report concerns and negative feedback, which informed refinements and are detailed below and in Table 4.
Psychoeducation Component
The PE component addressed stereotypes surrounding the behaviors, symptoms, and “appearance” of EDs. It aimed to validate users’ concerns and encourage participants to reflect on their personal eating and body image struggles. Users generally offered positive feedback on content addressing ED myths, as expressed by P3 in Cycle 1, “It was helpful…when you think [of] eating disorders, you think of people who are generally more skinny, that are throwing up or…something drastic.” Participants with BED responded most positively to this component, both pre- and post-refinements, as they felt the ED stereotype content normalized and validated their concerns. A common user suggestion included adding content about more complex ED presentations. A user in Cycle 2 (P9) reported, “I think it was incomplete because I’m struggling with several things right now and I just kind of had to pick one.” Other users suggested increasing personal relevance by enabling users to select multiple concerns (i.e., binge eating, restricting food, purging, and preoccupation or anxiety around food and body image) to learn more about.
To address suggestions, initial language refinements between iterations added more supportive and encouraging messages, improved conversation flow, and reduced self-blame. Content was revised to promote user reflection on information learned and to enable users to learn about multiple endorsed concerns.
In cycles following refinements, most users found the content adequately addressed a wide range of possible ED-related behaviors and challenges. One user in Cycle 3 (P14) noted, “…when you think of disordered [images] in the media it’s a lot of exercising and excessive purging and weight loss which isn’t really what I’m affected by. I'm affected by bingeing. It’s good to clarify misconceptions there too because that’s something you don’t see a lot anywhere else” (P18). Some users in later cycles still felt limited in their answer choices: “It’s like those phone menu things where it’s like ‘okay if you need this, press 0’… then you have to answer other questions” (P19). This highlights limitations inherent in a rule-based chatbot approach.
Motivational Interviewing Component
The MI component encouraged users to evaluate on a scale from 1-10 how important they felt it was to address their EDs and their confidence in making changes addressing related behaviors. Users were then asked to reflect on their responses. Participants generally reported these quantitative items fostered introspection on their actual confidence and motivation to change. P15 explained that the scales “…made me accountable at that time because I may say and think [about something] in my head, but when someone asked me a question and I saw the number in black and white in front of me, it made me accountable…It was basically my ‘aha’ moment.” Conversely, most negative feedback from users expressed that follow-up scale items (Table 4a) were confusing and difficult to respond to: “I didn’t mind doing the ranking but the follow-up questions where it asks why you pick a number, I didn’t know what would make my answer go from a 7 to an 8” (P4). Such follow-up scale items were revised between iterations, among other content and language changes, to promote greater self-reflection among users. Additional language revisions aimed to improve users’ motivation to pursue treatment and validate users with prior help-seeking challenges.
Participants responded well to the above refinements. They increasingly found that language in the module promoted honest disclosure and reflection, which augmented their motivation to work on their ED-related challenges. P10 found that the refined language “…encouraged me to be more honest and get a little more open instead of…minimizing.” Users also reflected more positively on revised follow-up questions (Table 4b-d), as opposed to early participants who received aforementioned scaled-follow-up items, and identified them as the primary driver of their increased motivation to seek treatment post-engagement.
Personalized Recommendation Componente
In the PR component, Alex prompted users to select seeing in-person or online help options before navigating users through options for each based on the decision trees illustrated in Figure 3. Users found the component easy to use, pragmatic, and introspective. Most positive feedback outlined that talking to Alex, versus a human, provided users with a pressure-free environment and decision-making autonomy. P19 explained, “…it was nice to have those options laid out…I can sit down and make my own decisions on what I need rather than someone telling me what I need.” While there were no concerns regarding navigation or ease of use in this component, users provided suggestions to improve their perceptions of provided recommendations. Participants wanted the choice to select both in-person and online resources, to decline help, and to see more resource options in the online help pathway. One user wanted to see more motivational language used to promote resource utilization. These suggestions informed refinements between iterations, which included adding validating language for users who faced prior barriers to seeking treatment and for users not yet ready for help. Content modifications added options for users to see both in-person and online resources and to receive multiple recommendations from each, to decline help, and to receive online self-help recommendations (Figure 3).
Following refinements, fewer users in Cycles 3 and 4 (2 of 11, 19%) provided suggestions for improvement than did users in Cycles 1 and 2 (6 of 10, 60%), and their feedback reinforced that content refinements improved personal salience of tailored recommendations to perceived user needs. Regarding the telehealth resources provided, P15 commented, “I loved it because I think this is the new normal…”. P17 explained, “Just knowing that there are more avenues besides…keeping this to yourself [or going] to one of those treatment centers. It's like, yeah, I can get on this whole online community. And, you know, that's a safe first step.”
Repeated Administration Component
The RA component was tested by users in Cycles 3 and 4, who received repeated follow-up check-ins over the two weeks following their initial conversation with Alex. Users generally found that the resource option reminders offered during these check-ins reinforced help-seeking behaviors. For example, P12 in Cycle 3 expressed, “It reminds you and you can’t just drop it. You get busy with things and forget to take care of the issue, but the chatbot reminded me.” Other users thought the check-ins solidified their motivation to seek treatment and provided reassurance that their disordered behaviors were “worthy” of care.
The primary concern users expressed in this component was the inability to schedule when check-in messages were received, including the day, time of day, and time between follow-up messages. The platform did not allow for a customized check-in schedule to be implemented.
Minor changes were made between Cycles 3 and 4, including content additions and language refinements to motivate help-seeking prior to the user’s next check-in.
Discussion
This paper describes the process of developing a user-informed chatbot, Alex, to increase motivation for treatment and mental health services use among a sample (N=21) screening positive for EDs and not yet in treatment. Feedback over four testing cycles guided modifications to each of four theoretically-informed components.
Feedback across cycles converged on four broad themes (user experience, chatbot qualities, chatbot content, and ease of use), which mirrored common themes in a recent review of 37 mental health chatbots exploring user experiences with conversational interventions (Abd-Alrazaq et al., 2021). Alex showed excellent usability and acceptance among participants even prior to refinements. Both in-person and remote participants reported no technical challenges or difficulty navigating the chatbot. Rather, suggestions for improvement addressed Alex’s content and language. This feedback was the basis of minor and major content refinements to the PE, MI, and PR components. Language refinements were included throughout all four components and enhanced Alex’s empathetic, supportive, and encouraging qualities. Over cycles of testing, fewer users reported finding Alex “robotic” or unaccommodating to their unique ED-related behaviors and resource needs. SUS scores demonstrated that perceived usability improved, albeit marginally, between iterations. However, usability did not improve linearly across cycles, likely attributable to the intervention being highly regarded from the first iteration.
Findings from other chatbot usability and acceptance studies suggest that perceived chatbot qualities (i.e., robotic, supportive, empathetic) are a likely mediator of general user acceptance and success of the intervention at improving mental health outcomes (Eisenstadt, Liverpool, Infanti, Ciuvat, & Carlsson, 2021). Similarly, users in our study largely expressed that Alex’s empathetic, personable, and human-like qualities augmented their honesty during conversations, trust in the resources and information provided, and confidence and willingness to take the next steps to address their ED and body image concerns. This phenomenon is known as the “digital therapeutic alliance”, which previous work deems critical to optimize chatbot utilization and user perceptions (Darcy, Daniels, Salinger, Wicks, & Robinson, 2021), and may explain the robust, positive feedback and usability scores received by Alex’s users across both in-person and remote testing environments.
Growing mental health resource demand has recently contributed to the development of more interventions using digital interfaces, yet very few integrate evidence-based content with user-informed refinements or test the applicability of these tools in remote settings (Torous et al, 2021). Other chatbot studies report users feeling inundated with resource options and feeling content was irrelevant to their mental health needs, leading to user disengagement and compromised intervention outcomes (Beilharz, Sukunesan, Rossell, Kulkarni, & Sharp, 2021). Alex was designed to bypass the aforementioned issues by 1) using theory-driven content iteratively refined based on users’ perceived needs and expectations; 2) providing users with a tailored experience, including individual specific resources and content; and 3) being optimized for use in clinical and remote settings.
To date, only three chatbots relating to EDs have emerged. The chatbot Tessa featured an established CBT-based EDs prevention program and aimed to improve body image and prevent EDs (Chan et al., 2022; Fitzsimmons-Craft et al., 2021). Results indicated that Tessa successfully reduced women’s concerns about weight and shape through 6-month follow-up and showed potential to reduce ED onset (Fitzsimmons-Craft et al., 2022). Another chatbot, KIT, was designed to support people with body image and eating concerns. KIT provided psychoeducation, information on help seeking, and coping skills to patients and their caregivers (Beilharz, Sukunesan, Rossell, Kulkarni, & Sharp, 2021). A chatbot similar to KIT is currently being developed and tested among adolescents in Brazil and aims to improve body image (Matheson et al., 2021). To our knowledge, Alex is the first chatbot designed for pairing with an EDs or other mental health-related online screen, with the goal of ultimately increasing service utilization. Uniquely within this population, perceived and internalized barriers, even above treatment access, may preclude patients from receiving or seeking treatment (Ali et al., 2017; Hamilton et al., 2022). As such, more interventions like Alex are needed that refer individuals to the diverse range of ED care options available (e.g., self-help apps) at little to no cost to patients and that can be accessed immediately. Greater development and demand of such tools that promote self-efficacy may grow awareness of hospitals and ED programs of innovative means to expand service options.
This study presents several limitations. Alex is a rule-based chatbot, which may introduce a comprehension barrier between itself and users. For example, some participants found the conversation robotic at times when Alex could not understand their more complex responses. Further development of Alex’s interface may include transforming it into an artificially intelligent chatbot, which uses natural language processing and is perceived as more “human-like” than rule-based chatbots (Boucher et al., 2021). Moreover, results may not be generalizable to some subgroups, given that study recruitment methods attracted individuals already receptive to new service models and motivated for care, and provided scarce representation of restrictive ED diagnoses, males, and non-white participants. Of note, a select few users felt their ED challenges were not severe enough to use Alex and expressed feelings of guilt for using the tool. Denial of illness severity remains one of the largest barriers to ED treatment, and these results suggest that digital interventions like Alex must be further adapted to combat such barriers. Future investigation is needed to 1) determine Alex’s efficacy at immediately and longitudinally improving motivation, self-efficacy, and help-seeking among users, including based on predictors like ED psychopathology and severity, and 2) ascertain which components of the chatbot are responsible for driving these improvements. A factorial study is currently underway to test these questions (Washington University School of Medicine, 2021).
Overall, this study provides preliminary evidence of the chatbot’s acceptability as a tool to improve motivation and help-seeking behaviors among individuals identified by an online screen as having an ED but not yet in treatment. If demonstrated efficacious for increasing services utilization among this population, Alex could be implemented broadly to increase uptake of treatment following screening for EDs, as well as adapted for pairing with screening for other mental health disorders.
Supplementary Material
Public Significance Statement:
Low rates of service utilization and treatment have been observed among individuals following online eating disorder screening. Tools are needed, including scalable, digital tools that can be easily paired with screening, to improve motivation for addressing eating disorders in order to promote service utilization.
Acknowledgements and Conflicts of Interest:
This research was supported by K08 MH120341 and R01 MH115128-04S1 from the National Institute of Mental Health and T32 HL130357 from the National Heart, Lung, and Blood Institute. The authors have no conflicts to declare.
Footnotes
Availability of Data, Materials, and Code: The data will be made available by reasonable request to the corresponding author.
References
- Abd-Alrazaq AA, Alajlani M, Ali N, Denecke K, Bewick BM, & Househ M (2021). Perceptions and Opinions of Patients About Mental Health Chatbots: Scoping Review. Journal of Medical Internet Research, 23, e17828. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Ali K, Farrer L, Fassnacht DB, Gulliver A, Bauer S, & Griffiths KM (2017). Perceived barriers and facilitators towards help-seeking for eating disorders: A systematic review. The International Journal of Eating Disorders, 50, 9–21. [DOI] [PubMed] [Google Scholar]
- Ali K, Fassnacht DB, Farrer L, Rieger E, Feldhege J, Moessner M, … Bauer S (2020). What prevents young adults from seeking help? Barriers toward help-seeking for eating disorder symptomatology. The International Journal of Eating Disorders, 53, 894–906. [DOI] [PubMed] [Google Scholar]
- American Psychiatric Association. (2006). Treatment of patients with eating disorders,third edition. American Psychiatric Association. The American Journal of Psychiatry, 163, 4–54. [PubMed] [Google Scholar]
- Bangor A, Kortum PT, & Miller JT (2008). An Empirical Evaluation of the System Usability Scale. International Journal of Human–Computer Interaction, 24, 574–594. [Google Scholar]
- Beilharz F, Sukunesan S, Rossell SL, Kulkarni J, & Sharp G (2021). Development of a Positive Body Image Chatbot (KIT) With Young People and Parents/Carers: Qualitative Focus Group Study. Journal of Medical Internet Research, 23, e27807. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Beintner I, Jacobi C, & Taylor CB (2012). Effects of an Internet-based prevention programme for eating disorders in the USA and Germany—A meta-analytic review. European Eating Disorders Review: The Journal of the Eating Disorders Association, 20, 1–8. [DOI] [PubMed] [Google Scholar]
- Bem DJ (1967). Self-perception: An alternative interpretation of cognitive dissonance phenomena. Psychological Review, 74, 183–200. [DOI] [PubMed] [Google Scholar]
- Boucher EM, Harake NR, Ward HE, Stoeckl SE, Vargas J, Minkel J, … Zilca R (2021). Artificially intelligent chatbots in digital mental health interventions: A review. Expert Review of Medical Devices, 18, 37–49. [DOI] [PubMed] [Google Scholar]
- Braun V, & Clarke V (2006). Using thematic analysis in psychology. Qualitative Research in Psychology, 3, 77–101. [Google Scholar]
- Brooke J (1995). SUS: A quick and dirty usability scale. Usability Eval. Ind, 189. [Google Scholar]
- Buntine W (1992). Learning classification trees. Statistics and Computing, 2, 63–73. [Google Scholar]
- Burke BL, Arkowitz H, & Menchola M (2003). The efficacy of motivational interviewing: A meta-analysis of controlled clinical trials. Journal of Consulting and Clinical Psychology, 71, 843–861. [DOI] [PubMed] [Google Scholar]
- Cameron G, Cameron D, Megaw G, Bond R, Mulvenna M, O’Neill S, … McTear M (2019). Assessing the Usability of a Chatbot for Mental Health Care. In Bodrunova SS, Koltsova O, Følstad A, Halpin H, Kolozaridi P, Yuldashev L, … Niedermayer H (Eds.), Internet Science (pp. 121–132). Cham: Springer International Publishing. [Google Scholar]
- Car LT, Dhinagaran DA, Kyaw BM, Kowatsch T, Joty S, Theng Y-L, & Atun R (2020). Conversational Agents in Health Care: Scoping Review and Conceptual Analysis. Journal of Medical Internet Research, 22, e17158. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Chan WW, Fitzsimmons-Craft EE, Smith AC, Firebaugh M-L, Fowler LA, DePietro B, … Jacobson NC (2022). The Challenges in Designing a Prevention Chatbot for Eating Disorders: Observational Study. JMIR Formative Research, 6, e28003. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Collins LM (2018). Conceptual Introduction to the Multiphase Optimization Strategy (MOST). In Collins LM (Ed.), Optimization of Behavioral, Biobehavioral, and Biomedical Interventions: The Multiphase Optimization Strategy (MOST) (pp. 1–34). Cham: Springer International Publishing. [Google Scholar]
- Copeland L, McNamara R, Kelson M, & Simpson S (2015). Mechanisms of change within motivational interviewing in relation to health behaviors outcomes: A systematic review. Patient Education and Counseling, 98, 401–411. [DOI] [PubMed] [Google Scholar]
- Creswell JW, & Creswell JD (2017). Research Design: Qualitative, Quantitative, and Mixed Methods Approaches. SAGE Publications. [Google Scholar]
- Darcy A, Daniels J, Salinger D, Wicks P, & Robinson A (2021). Evidence of Human-Level Bonds Established With a Digital Conversational Agent: Cross-sectional, Retrospective Observational Study. JMIR Formative Research, 5, e27868. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Denecke K, Abd-Alrazaq A, & Househ M (2021). Artificial Intelligence for Chatbots in Mental Health: Opportunities and Challenges. Multiple Perspectives on Artificial Intelligence in Healthcare, 115–128. [Google Scholar]
- Eisenberg D, Nicklett EJ, Roeder K, & Kirz NE (2011). Eating Disorder Symptoms Among College Students: Prevalence, Persistence, Correlates, and Treatment-Seeking. Journal of American College Health, 59, 700–707. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Eisenstadt M, Liverpool S, Infanti E, Ciuvat RM, & Carlsson C (2021). Mobile Apps That Promote Emotion Regulation, Positive Mental Health, and Well-being in the General Population: Systematic Review and Meta-analysis. JMIR Mental Health, 8, e31170. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Fairburn CG, & Beglin SJ (1994). Assessment of eating disorders: Interview or self-report questionnaire? The International Journal of Eating Disorders, 16, 363–370. [PubMed] [Google Scholar]
- Fenerty SD, West C, Davis SA, Kaplan SG, & Feldman SR (2012). The effect of reminder systems on patients’ adherence to treatment. Patient Preference and Adherence, 6, 127–135. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Festinger L (1962). A Theory of Cognitive Dissonance. Stanford University Press. [Google Scholar]
- Fitzsimmons-Craft EE, Balantekin KN, Graham AK, Smolar L, Park D, Mysko C, … Wilfley DE (2019). Results of disseminating an online screen for eating disorders cross the U.S.: Reach, respondent characteristics, and unmet treatment need. International Journal of Eating Disorders, 52, 721–729 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Fitzsimmons-Craft EE, Balantekin KN, Graham AK, DePietro B, Laing O, Firebaugh ML, … Wilfley DE (2020). Preliminary data on help-seeking intentions and behaviors of individuals completing a widely available online screen for eating disorders in the United States. International Journal of Eating Disorders, 53, 1556–1562. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Fitzsimmons-Craft EE, Chan WW, Smith AC, Firebaugh M-L, Fowler LA, Topooco N, … Jacobson NC (2022). Effectiveness of a chatbot for eating disorders prevention: A randomized clinical trial. International Journal of Eating Disorders, 55, 343–353. [DOI] [PubMed] [Google Scholar]
- Fulmer R, Joerin A, Gentile B, Lakerink L, & Rauws M (2018). Using Psychological Artificial Intelligence (Tess) to Relieve Symptoms of Depression and Anxiety: Randomized Controlled Trial. JMIR Mental Health, 5, e9782. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Graham AK, Trockel M, Weisman H, Fitzsimmons-Craft EE, Balantekin KN, Wilfley DE, & Taylor CB (2019). A screening tool for detecting eating disorder risk and diagnostic symptoms among college-age women. Journal of American College Health: J of ACH, 67, 357–366. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Graham AK, Wildes JE, Reddy M, Munson SA, Barr Taylor C, & Mohr DC (2019). User-centered design for technology-enabled services for eating disorders. The International Journal of Eating Disorders, 52, 1095–1107. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Gratzer D, Torous J, Lam RW, Patten SB, Kutcher S, Chan S, … Yatham LN (2021). Our Digital Moment: Innovations and Opportunities in Digital Mental Health Care. The Canadian Journal of Psychiatry, 66, 5–8. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Hamilton A, Mitchison D, Basten C, Byrne S, Goldstein M, Hay P, … Touyz S (2022). Understanding treatment delay: Perceived barriers preventing treatment-seeking for eating disorders. Australian & New Zealand Journal of Psychiatry, 56, 248–259. [DOI] [PubMed] [Google Scholar]
- Hannah DR, & Lautsch BA (2011). Counting in Qualitative Research: Why to Conduct it, When to Avoid it, and When to Closet it. Journal of Management Inquiry, 20, 14–22. [Google Scholar]
- Head KJ, Noar SM, Iannarino NT, & Grant Harrington N (2013). Efficacy of text messaging-based interventions for health promotion: A meta-analysis. Social Science & Medicine (1982), 97, 41–48. [DOI] [PubMed] [Google Scholar]
- Hennink MM, Kaiser BN, & Marconi VC (2017). Code Saturation Versus Meaning Saturation: How Many Interviews Are Enough? Qualitative Health Research, 27, 591–608. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Islam MN, Khan S, Islam N, Rownok R, Zaman R, & Zaman S (2021). A Mobile Application for Mental Health Care During COVID-19 Pandemic: Development and Usability Evaluation with System Usability Scale. [Google Scholar]
- Jaspers MWM, Steen T, van den Bos C, & Geenen M (2004). The think aloud method: A guide to user interface design. International Journal of Medical Informatics, 73, 781–795. [DOI] [PubMed] [Google Scholar]
- Kazdin AE, Fitzsimmons-Craft EE, & Wilfley DE (2017). Addressing critical gaps in the treatment of eating disorders. The International Journal of Eating Disorders, 50, 170–189. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Kreuter MW, & Wray RJ (2003). Tailored and targeted health communication: Strategies for enhancing information relevance. American Journal of Health Behavior, 27 Suppl 3, S227–232. [DOI] [PubMed] [Google Scholar]
- Lewis JR, & Sauro J (2009). The Factor Structure of the System Usability Scale. In M. Kurosu. The effectiveness and applicability of motivational interviewing: A practice-friendly review of four meta-analyses. Journal of Clinical Psychology, 65, 1232–1245. [DOI] [PubMed] [Google Scholar]
- Lund A (2001). Measuring Usability with the USE Questionnaire. Usability and User Experience Newsletter of the STC Usability SIG, 8. [Google Scholar]
- Lundahl B, & Burke BL (2009). The effectiveness and applicability of motivational interviewing: A practice-friendly review of four meta-analyses. Journal of Clinical Psychology, 65, 1232–1245. [DOI] [PubMed] [Google Scholar]
- Lyon AR, & Bruns EJ (2019). User-Centered Redesign of Evidence-Based Psychosocial Interventions to Enhance Implementation-Hospitable Soil or Better Seeds? JAMA Psychiatry, 76, 3–4. [DOI] [PubMed] [Google Scholar]
- Matheson EL, Smith HG, Amaral ACS, Meireles JFF, Almeida MC, Mora G, … Diedrichs PC (2021). Improving body image at scale among Brazilian adolescents: Study protocol for the co-creation and randomised trial evaluation of a chatbot intervention. BMC Public Health, 21, 2135. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Miller WR, & Rollnick S (2012). Motivational Interviewing: Helping People Change. Guilford Press. [Google Scholar]
- Misono AS, Cutrona SL, Choudhry NK, Fischer MA, Stedman MR, Liberman JN, … Shrank WH (2010). Healthcare information technology interventions to improve cardiovascular and diabetes medication adherence. The American Journal of Managed Care, 16, SP82–92. [PubMed] [Google Scholar]
- Molina-Recio G, Molina-Luque R, Jiménez-García AM, Ventura-Puertos PE, Hernández-Reyes A, & Romero-Saldaña M (2020). Proposal for the User-Centered Design Approach for Health Apps Based on Successful Experiences: Integrative Review. JMIR MHealth and UHealth, 8, e14376. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Montenegro JLZ, da Costa CA, & da Rosa Righi R (2019). Survey of conversational agents in health. Expert Systems with Applications, 129, 56–67. [Google Scholar]
- Noar SM, Benac CN, & Harris MS (2007). Does tailoring matter? Meta-analytic review of tailored print health behavior change interventions. Psychological Bulletin, 133, 673–693. [DOI] [PubMed] [Google Scholar]
- Oh J, Jang S, Kim H, & Kim J-J (2020). Efficacy of mobile app-based interactive cognitive behavioral therapy using a chatbot for panic disorder. International Journal of Medical Informatics, 140, 104171. [DOI] [PubMed] [Google Scholar]
- Pereira J, & Díaz Ó (2019). Using Health Chatbots for Behavior Change: A Mapping Study. Journal of Medical Systems, 43, 1–13. [DOI] [PubMed] [Google Scholar]
- Petty RE, Cacioppo JT, & Goldman R (1981). Personal involvement as a determinant of argument-based persuasion. Journal of Personality and Social Psychology, 41, 847–855. [Google Scholar]
- Sauro J (2011). A Practical Guide to the System Usability Scale: Background, Benchmarks & Best Practices. Measuring Usability LLC. [Google Scholar]
- Schaumberg K, Welch E, Breithaupt L, Hübel C, Baker JH, Munn-Chernoff MA, … Bulik CM (2017). The Science Behind the Academy for Eating Disorders’ Nine Truths About Eating Disorders. European Eating Disorders Review : The Journal of the Eating Disorders Association, 25, 432–450. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Stephens TN, Joerin A, Rauws M, & Werk LN (2019). Feasibility of pediatric obesity and prediabetes treatment support through Tess, the AI behavioral coaching chatbot. Translational Behavioral Medicine, 9, 440–447. [DOI] [PubMed] [Google Scholar]
- Taylor CB, Bryson S, Luce KH, Cunning D, Doyle AC, Abascal LB, … Wilfley DE (2006). Prevention of eating disorders in at-risk college-age women. Archives of General Psychiatry,63,881–888. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Taylor CB, Kass AE, Trockel M, Cunning D, Weisman H, Bailey J, … Wilfley DE (2016). Reducing Eating Disorder Onset in a Very High Risk Sample with Significant Comorbid Depression: A Randomized Controlled Trial. Journal of Consulting and Clinical Psychology, 84, 402–414. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Thakkar J, Kurup R, Laba T-L, Santo K, Thiagalingam A, Rodgers A, … Chow CK (2016). Mobile Telephone Text Messaging for Medication Adherence in Chronic Disease: A Meta-analysis. JAMA Internal Medicine, 176, 340–349. [DOI] [PubMed] [Google Scholar]
- Torous J, Bucci S, Bell IH, Kessing LV, Faurholt-Jepsen M, Whelan P, … Firth J (2021). The growing field of digital psychiatry: Current evidence and the future of apps, social media, chatbots, and virtual reality. World Psychiatry, 20, 318–335. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Torous J, & Haim A (2018). Dichotomies in the Development and Implementation of Digital Mental Health Tools. Psychiatric Services, 69, 1204–1206. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Washington University School of Medicine. (2021). Developing an Optimized Conversational Agent or “Chatbot” to Facilitate Mental Health Services Use in Individuals With Eating Disorders (Clinical Trial Registration No. NCT04806165). Clinicaltrials.gov.
- Wozney LM, Baxter P, Fast H, Cleghorn L, Hundert AS, & Newton AS (2016). Sociotechnical Human Factors Involved in Remote Online Usability Testing of Two eHealth Interventions. JMIR Human Factors, 3, e4602. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Zhang J, Oh YJ, Lange P, Yu Z, & Fukuoka Y (2020). Artificial Intelligence Chatbot Behavior Change Model for Designing Artificial Intelligence Chatbots to Promote Physical Activity and a Healthy Diet: Viewpoint. Journal of Medical Internet Research, 22, e22845. [DOI] [PMC free article] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.