Abstract
This review reports the user experience of symptom checkers, aiming to characterize users studied in the existing literature, identify the aspects of user experience of symptom checkers that have been studied, and offer design suggestions. Our literature search resulted in 31 publications. We found that (1) most symptom checker users are relatively young; (2) eight relevant aspects of user experience have been explored, including motivation, trust, acceptability, satisfaction, accuracy, usability, safety/security, and functionality; (3) future symptom checkers should improve their accuracy, safety, and usability. Although many facets of user experience have been explored, methodological challenges exist and some important aspects of user experience remain understudied. Further research should be conducted to explore users’ needs and the context of use. More qualitative and mixed-method studies are needed to understand actual users’ experiences in the future.
Introduction
The rapid development of digital technology and the healthcare industry has contributed to the increased popularity of symptom checkers (SCs) in application (app) markets. Some SC apps (e.g., Ada) have been downloaded from app stores tens of millions of times1. These SCs utilize either a chatbot or a questionnaire-like form to solicit symptom input and provide users with a potential diagnosis. Some SCs describe themselves as akin to intelligent doctors who give accurate and reliable medical recommendations. For instance, Ada claims to “think like a doctor”2. However, the reliability of SCs is questionable3. Healthcare consumers may put their health at risk if they uncritically accept the diagnoses from SCs4. Recognizing that individuals’ perceptions and interpretations of SCs’ responses may influence their subsequent actions, it is important to understand more fully the user experience (UX) of SCs5. UX is “a consequence of a user’s internal state, the characteristics of the designed system, and the context (or the environment) within which the interaction occurs.6” As Table 1 indicates, we adapted this definition of UX to characterize the UX of SCs along three dimensions. As the UX of SCs is highly related to the acceptance of SCs7 and the minimization of unintended consequences (e.g., anxiety and cyberchondria)8, it is crucial to improve the UX of SCs by understanding individuals’ motivations for using SCs, users’ expectations of SCs, the benefits as well as the limitations of SCs’ features, and the usage contexts.
Table 1.
Three dimensions of the UX of SCs.
Dimension | Definition |
---|---|
Consequence of a user’s internal state | User’s expectations of, needs for, and attitude towards SCs |
Characteristics of the designed system | SCs’ features, including usability, complexity, purpose, and functionality |
Context or environment | Context or environment within which users interact with SCs |
Nevertheless, few existing papers have systematically reviewed studies of the UX of SCs. Instead, previous systematic reviews of SCs have focused on the influence of SCs’ use on the patient–doctor relationship9, the current evaluation methodologies for SCs1, and the impact of SCs on urgent care seeking10 and primary care11. To the best of our best knowledge, only one scoping review of SCs covered studies addressing the UX of SCs12. This scoping review provided a broad view of research on SCs and points out that UX is one theme that has been explored. However, it did not offer an in-depth analysis of the aspects of the UX of SCs that have been explored. It is still unclear who SC users are, what factors motivate and influence the use of SCs, and what relevant design suggestions might be. To address this gap in existing reviews, we systematically reviewed papers that examine the UX of SCs. We choose the systematic review approach instead of doing a scoping review because a systematic review can allow us to synthesize the findings from previous literature and identify areas for future research13. We intended to address the following three questions: (1) Who uses SCs? (2) Which aspects of the UX of SCs have been studied based on the three dimensions of UX? and (3) What design suggestions have been proposed to improve the UX of SCs?
Materials and Methods
Search strategy and screening process. We developed a study protocol in compliance with the PRISMA guidelines. The PROSPERO website was searched and no previous systematic review on similar topics was found, which validated the novelty of our review. We searched six electronic databases from May 2021 to July 2021: PubMed, CINAHL, ACM Digital Library, Scopus, Web of Science, and Google Scholar. We also conducted backward–forward searches by examining the references from relevant articles and systematic reviews. We utilized three groups of search terms: SC-related terms, internet technology-related terms, and health-related terms. The following search query was used for Scopus: (TITLE-ABS-KEY(“symptom checker” OR “self diagnosis” OR “self-diagnosis” OR “self triage” OR “self-assessment” OR “self assessment” OR “self diagnosing” OR “self-evaluation” OR “self evaluation” OR “self-appraisal” OR “check your symptom” OR “check their symptom” OR “online diagnosis” OR “web-based triage” OR “electronic triage” OR “etriage” OR “self referral” OR “assistance technology” OR “assistance technologies”) AND TITLE-ABS-KEY(“online” OR “technology” OR “website” OR “tool” OR “computer” OR “mhealth” OR “m-health” OR “ehealth” OR “ e-health” OR “app” OR “mobile application” OR “smartphone” OR “smart phone” OR “cell phone” OR “cellular phone” OR “mobile phone” OR “electronic” OR “automated” OR “internet” OR “digital” OR “mobile”) AND TITLE-ABS-KEY(“medical” OR “health” OR “medication” OR “healthcare”)). The search strings for Scopus were adapted for other databases.
After completing the database and backward–forward searches, we conducted two rounds of screening to determine which articles to include in our review. First, we screened the titles and abstracts of the identified papers using three eligibility criteria: (1) the paper is an original research article published in a peer-reviewed journal or conference proceedings; (2) the paper was published in English; and (3) the paper’s title and abstract included our search terms. We then screened the full texts of those papers based on the following inclusion criteria: (1) the subject of the paper is an application rather than a model or algorithm or general online search; (2) consumers/laypersons are the primary users of the technology developed and/or evaluated in the paper; (3) the technology described in the paper is internet-enabled; (4) the technology studied in the paper is used for self-diagnosis; (5) the topic of the paper addresses an aspect of the UX of SCs; and (6) the full text is available. After each round of screening, three researchers performed a cross-review and settled any disagreements. Only those papers that the researchers deemed sufficiently relevant were included. In addition, quality assessment was conducted for all papers to assess whether these papers’ aims and methodology were clearly described, whether their findings align with their research questions, and whether they were peer-reviewed. All discrepancies were discussed and resolved. The quality assessment results were taken into account when synthesizing data. We finally identified 16,825 potentially relevant articles written in the English language. We also identified 3,425 papers using Google Scholar and backward–forward searching. Ultimately, 31 papers (24 from databases and 7 from other methods) were chosen to be included. Figure 1 shows the screening process.
Figure 1.
Processes of identification and screening.
Data collection and analysis. Using the papers identified during the two rounds of screening, we conducted data extraction based on the PRISMA guidelines. We collected the following main information: (1) the source of the paper, including the author(s), date of publication, and publication venue; (2) background information, including the diseases examined in the paper, the research purposes, stakeholders, and the features/functions of SCs; (3) methodological information, including each author’s country and affiliation, the study setting and duration, the participants’ characteristics, the data collection and analysis methods; and (4) findings, including the participants’ motivations for using SCs, the participants’ characteristics, the dimensions/measurements of UX, and design suggestions. We then used thematic analysis and descriptive statistical analysis to synthesize the extracted data.
Results
Study characteristics. The 31 papers chosen for inclusion were published between 2007 and 2021 (see Table 2).
Table 2.
List of studies included in the systematic review.
Source | Country | Study Design | Methodology | Diseases | Participants | System Features |
---|---|---|---|---|---|---|
Powley et al, 201614 | UK | Laboratories study | Quantitative; Experimental study | Inflammatory joint disease | N = 34; Age: 29.75-78.5; Female | Existing apps; Questionnaire-based |
Bisson et al, 201615 | US | Laboratories study | Quantitative; Experimental study | Knee pain | N = 328; Age: 18-76; 49.7% male | Prototype; Questionnaire-based |
Bisson et al, 201416 | US | Field study | Quantitative; Experimental study | Knee pain | N = 527; Age: avg: 47, 18-84; 51.6% male | Prototype |
Hageman et al, 201517 | US | Laboratories study | Quantitative; Experimental study & questionnaire | Hand diseases | N = 86 patients; Avg age: 46; 51% male | Existing apps; Questionnaire-based |
Knitza et al, 202118 | Germany | Laboratories study | Quantitative; Experimental study | Inflammatory rheumatic disease | N = 164 patients | Existing apps; Chatbot & questionnaire-based; AI-driven & rule-based |
Cross et al, 202119 | Australia | Laboratories study | Quantitative; Experimental study & questionnaire & log-based study | N/A | N = 64; Age: > 18, 87.49%: 18-34; 4.68%: 35-44; 1.56%: 45-54 | A clone of existing app; Questionnaire-based |
Verzantvoort et al, 201820 | the Netherlands | Field study | Mixed methods; Experimental study & interview | Acute primary care | N = 4,456; Age: most 19-45 (56.4%) | Existing app |
Fan et al, 202121 | China | Field study | Mixed methods; Log-based study | N/A | N = 16,519; Age: most 20-39; 54.8% male | Existing app; Chatbot; AI-driven |
Price et al, 201322 | US | Laboratories study | Quantitative; Experimental study & questionnaire | Influenza-Like Illness | N = 294 parents and caregivers; Age: > 18 | Prototype; Questionnaire-based |
You & Gui, 202023 | US | Field study | Qualitative; App review analysis & interview | N/A | N = 10 (interview); Age: 23-40 (interview) | Existing apps; Chatbot |
Aboueid et al, 20217 | Canada | Laboratories study | Qualitative; Think aloud & interview | N/A | N = 24; Age: 18-34 | Existing apps; Questionnaire-based & chatbot |
Aboueid et al, 202124 | Canada | Laboratories study | Qualitative; Interview | COVID-19 | N = 22; Age: 18–33; 54.5% female | Existing apps; Questionnaire-based & chatbot |
Marco-Ruiz et al, 201725 | Norway | Laboratories studies | Mixed methods; Experimental study & think aloud | Respiratory diseases | Phase 1: N = 53; Phase 2: N = 15 | Existing apps |
Ponnada, 202026 | US | Laboratories study | Mixed methods; Usability testing & interview | COVID-19 | N/A | Existing app & improved prototype; Chatbot |
Schrager et al, 202027 | US | Laboratories study | Quantitative; Prototyping & usability testing | COVID-19 | N = 877; Age: avg: 32, 19-84; 65.3% female | Prototype; Questionnaire-based |
Reilly & Austin, 202128 | Australia | Field study | Quantitative; Online survey | Depression in pregnancy | N = 140; Age: N/A; Female | Existing app; Questionnaire-based |
Nieroda et al, 201829 | UK | Laboratories study | Qualitative; Think aloud & online survey | Cancer | The public: N=39, Age: >= 40; Practitioners: N = 20 | Existing app |
Daher et al, 202030 | Switzerland | Laboratories study | Quantitative; Experimental study | N/A | N = 12; Age: N/A | Prototype; Chatbot |
Tsai et al, 202131 | US | Field study & laboratories study | Mixed methods; Interview & experimental study | General/ COVID-19 | (1) N = 25; Age: 19-54;(2): N = 20; Age: 19-37; 70% female | Existing apps & prototype; Chatbot & questionnaire-based |
Li et al, 202032 | China | Field study & laboratories study | Mixed methods; Interview & experimental study | N/A | (1) N = 13; Age: 21-61 (2) N = 48 | Existing apps; Chatbot; AI-driven |
You et al, 202133 | China | Field study | Qualitative; Interview | N/A | N = 30; Age: 20-55 Most were in their 20s and 30s | Existing apps; Chatbot & questionnaire-based; AI-driven |
Hwang et al, 201834 | US, South Korea | Laboratories study | Qualitative; Think aloud | N/A | N = 6; Age: 20-39; 4 female, 2 male | Existing apps; Chatbot; AI-driven |
Baldauf et al, 202035 | Switzerland | Questionnaires | Mixed methods; Online survey | N/A | N = 106; Age: avg: 34, 19-60; 58.5% male | Existing apps; Using sensor data; AI-driven |
Miller et al, 202036 | UK | Laboratories study | Quantitative; Experimental study & questionnaire | N/A | N = 523; Age: avg: 39.79, 81.6%: 15-64; 62.1% female | Existing app Chatbot; AI-driven |
Meyer et al, 202037 | US | Questionnaires | Mixed methods; Online survey | N/A | N = 329; Age: avg: 48.0; 75.7% female | Existing app; questionnaire,AI-based |
Winn et al, 201938 | US | Field study | Quantitative; Log analysis | N/A | N = 158,083; Age: avg: 40; 78% female | Existing app; Chatbot |
DeForte et al, 202039 | US | Field study | Qualitative; App review analysis | Depression | N/A | Existing apps |
Nijland et al, 201040 | the Netherlands | Field study & questionnaires | Quantitative; Log-based & survey | N/A | N = 192; Age: 16–35 (56%); 66% female | Existing app; Questionnaire-based |
Luger et al, 201441 | US | Laboratories study | Qualitative; Think aloud&questionnaire | N/A | N = 79; Age: >= 50; | Existing app; Questionnaire-based |
Lanseng et al, 200742 | Norway | Questionnaires | Quantitative; Online survey | N/A | N = 160; Age: 18-65; 46.3% female | N/A |
Hua & Hou, 202043 | Vietnam | Field study & questionnaires | Mixed methods; Interview & survey | N/A | N = 482; Age: 18-50; 63.1% female | N/A |
Characteristics of users. Four studies that we reviewed (12.9%)7,21,37,38 reported on the characteristics of individuals who had used SCs prior to participating in the studies. The age of SC users was found to be relatively young (most users ranged in age from 20 to 39 years old21 or the average age was 40 years old38). Regarding the gender of SC users, the study conducted in China found male users initiated most consultations (54.80%)21, while the U.S. study reported most users were female (75.7%)37. Other less commonly mentioned user characteristics in our corpus included health literacy7, education37, and income37. Young individuals with low health literacy and high technology literacy were found to be inclined to use SCs7. The household income of most users was less than $100,000 USD37.
UX aspects of SCs.
(1) Consequence of a consumer’s internal state
Motivations for using SCs: Five studies (16.1%)21,28,33,38,43 addressed users’ motivations for using SCs (see Table 3).
Table 3.
Motivations for using SCs (n = 5).
Motivation | Details |
Convenience | The researchers found that individuals were motivated to use SCs instead of going to a doctor because of the time and cost savings as well as location independence33 |
Supplementing offline medical visits | Individuals were motivated to use SCs to supplement in-person medical visits and gain assistance in making healthcare-related decisions28,33,38 |
Gaming | Individuals (8.03%) gamed SC by entering nonsensical words for nontherapeutic purposes21,43 |
Social influence | Social influence (i.e., “the extent to which an individual is aware that significant others recommend they use the self-diagnostic app”) could prompt consumers to use SCs43 |
Trust: Six of the reviewed papers (19.4%) explored users’ trust in SCs21,31–35, reporting factors that may influence users’ trust, including the provision of explanations, interaction design, and accuracy (see Table 4).
Table 4.
Factors that influence the users’ trust (n = 6).
Factor | Impact | Methods |
---|---|---|
Explanations | The information source or authority31,33, medical certification35, security of data transmission35, and explanations for presented questions and reports31,34 could positively influence users’ trust | Interviews33, think aloud34, questionnaires (adapted from previous literature31, self-created35) |
Interaction design | Repeated or irrelevant questions34, vague options34, limited input functions33, and an overly anthropomorphic conversation design33 could negatively affect users’ trust in SCs | Interviews33, think aloud34 |
Accuracy | Users tended to distrust SCs when the SCs’ recommendations were inaccurate21,32,33 | Interviews33, log analysis21, questionnaires (self-created)32 |
Acceptability: In our corpus, a total of four papers (12.9%) measured whether users accepted SCs18,28,29,36. These studies relied on questionnaires18,28,29,36 and the thinking-aloud method29. All of the papers described how most users accepted the use of SCs to support emotional wellness28, primary care36, and cancer diagnosis29.
Satisfaction: Four of the studies (12.9%) evaluated individuals’ satisfaction with using SCs via questionnaires19–21,31. These studies found that some of the relevant factors affecting satisfaction are disease type21, accuracy21, and explanations31. One log-based study found that for common diseases, users were usually satisfied with the diagnostic recommendations, while for more complex diseases, users were less satisfied21. This study also reported that inaccurate diagnostic results contributed to dissatisfaction with SCs. One experimental study reported that providing different styles of explanations can increase users’ satisfaction with SCs31.
(2) Characteristics of the SCs
Accuracy: Eleven papers included in our corpus examined the accuracy of SCs. Three papers stated that compared with physicians’ diagnoses, recommendations from SCs were relatively accurate in diagnosing knee pain (correctly diagnose 58% of the time)15, acute primary care (correctly triage 89% of patients)20, and influenza-like illness (correctly diagnose 93% of the time)22, though two of them admitted that SCs had low specificity16,22. However, the other eight papers revealed that SCs are frequently inaccurate when providing diagnoses for inflammatory joint diseases14, hand diseases17, inflammatory rheumatic diseases18, knee pain15, and general diseases (e.g., 1,084 out of 3,832 diagnoses were rated by users as inaccurate21)19,21,37,41.
Usability: Ten papers explored the usability of SCs. Three experiment-based papers reported that SCs were easy to use18,22,36 (e.g., 97.8% of the participants considered Ada easy to use36) by using the standardized scales (e.g., System Usability Scale)18,36 or self-created web survey22. The remaining seven studies reported usability issues (see Table 5).
Table 5.
Usability issues (n = 10).
Usability issues | Details | Methods |
---|---|---|
Neglect of user groups and diseases | SCs could not support diverse user groups (e.g., transgender individuals) or diseases23 | Interviews23, app review analysis23 |
Input limitations | Users perceived difficulties in inputting the information of their diseases completely (62%)20, inputting sufficient patient history information23, describing disease dimensions (e.g., frequency)23,44, and entering medical terms that could be recognized by the SCs23 | Questionnaires (self-created)20, think aloud44, interviews23, app review analysis23 |
Problematic questions | Users perceived that the probing questions of SCs were seemingly irrelevant based and were presented in an unreasonable sequence23 | Interviews23, app review analysis23 |
Interpretation issues | Users encountered challenges in understanding medical jargon7, complex questions23, and diagnostic reports21,31 | Interviews7,23, app review analysis23, log analysis21, questionnaires (adapted from previous literature)31 |
Overwhelming amounts of information | Users complained of too much information on the welcome page44, excessive subsections to fill in44, too many pieces of information presented at once26, and redundant questions23 | App review analysis23, interviews23,26, usability testing26, think aloud44 |
Safety/security: Three studies highlighted that using SCs may have potential risks for both the healthcare system and consumers22,28,33. For instance, the specificity of the SC used for influenza-like illness is poor, which may lead more users to go to the emergency department, thereby putting pressure on healthcare systems22. Likewise, since most health consumers are laypersons who may not have sufficient medical knowledge, they cannot discern the correctness of SCs’ recommendations. Using SCs uncritically may compromise users’ health and safety28,33. For example, one paper found only 53% to 61% of users with depression reported by the focal SC discussed their depression with medical professionals28. This might result in untreated depression. Two papers emphasized the significance of data security35,44, calling for more information about how SCs encrypt35, anonymize35, and store data44.
Functionality: Two studies examined the functions of SCs using an app feature analysis23, an app review analysis23,39, and semi-structured interviews23. The first study examined 11 apps released in the United States via the Apple App store and the Google Play store23. The researchers found most apps support five procedures: creating a patient history for users, evaluating symptom input by users, providing recommendations, ordering further tests, and providing follow-up treatment (e.g., contacting healthcare practitioners). The other study showcased the multiple functions users desired, such as mood tracking, journaling, and providing educational materials39.
(3) Context
Four studies20,21,38,40 mentioned the context within which consumers interacted with SCs. They reported that SCs were used to address myriad symptoms. Two of them reported users mostly used SCs to address non-emergency care issues, such as common cold symptoms40, gastroenterology21, and dermatology21, while fewer individuals used SCs in emergent situations21. The SC DoctorBot was also used for diseases that may carry social stigma21. One of the studies found that the most common symptoms reported by users are pain and abdominal functioning38.
Design suggestions for improving the UX of SCs. Based on their findings, the papers included in our corpus offered design recommendations for improving the UX of SCs. The details are presented in Table 6.
Table 6.
Design suggestions.
Aspects | Design suggestions |
---|---|
Accuracy | Increase the accuracy of SCs by considering the probability of specific conditions17 and the uniqueness of a variety of symptom clusters17 |
Safety | |
Functions | |
Broader healthcare information |
|
Usability |
|
Discussion
This systematic review is the first to address the UX of SCs. We reported on and analyzed the characteristics of the published literature, SC users’ characteristics, the motivations of individuals to use SCs, the relevant aspects of the UX of SCs, and design implications. In this section, we discuss the implications of our results for future research.
Overall, the UX studies in our corpus focus primarily on the characteristics of the SCs rather than investigating consumers’ internal states. The included studies put more emphasis on evaluating rather than designing SCs centering around users’ needs. In healthcare, patient-centered care (i.e., “providing care that is respectful of and responsive to individual patient preferences, needs, and values”45) is always highlighted. When designing healthcare technologies, users’ expectations and needs should also be centered and considered46. However, insufficient attention has been paid to fully understanding users’ needs for SCs, especially their concerns and expectations in everyday practices. Our findings show that most studies have explored the use of SCs in laboratory or otherwise controlled settings (n=16). While we acknowledge the importance of laboratory experiments, we suggest that studies that are deployed in the wild are needed to capture individuals’ unanticipated use of SCs and reveal UX issues that only arise in-situ47.
In addition, patient-centered care suggests giving respect, showing emotional support, communication skills, and education, and offering explanations48,49. However, whether and how this line of guidelines should be applied to SC design is still unclear. First, our systematic review found that few studies focus on how to provide emotional support to SC users and improve their hedonic experience. Only four studies highlighted that SCs should convey human empathy26,30,32. Little is known about how to deliver emotional support to users. Moreover, although existing healthcare research suggests that humanoid features of healthcare technologies may improve communication and therefore UX50,51, little research in our corpus put forward recommendations for how to embed humanoid features in SCs to improve the hedonic experience. Further research should be conducted to investigate what level of empathy should be provided and how the interface design of SCs can be improved. Second, though four studies we reviewed investigated the explanations21,31,34,35, which emphasized the importance of providing explanations (e.g., explanations for presented questions and diagnostic reports31,34) to improve the transparency of SCs and users’ trust, it is still unclear how we should deliver explanations to users. Future research might explore how to give users more control and customization over the offerings of explanations31.
Furthermore, few studies in our corpus focus on the context of use. Existing research has highlighted the significance of the context in which technology is embedded52, recognizing that technology is integrated into everyday practices52. Health information needs are especially likely to vary across locations53. It is crucial to consider the geographic and other situated contexts within which the user interactions with SC happen. For instance, our review shows that most papers on SCs are Western in origin and their studies were conducted in developed countries. Few studies were conducted in developing countries such as Vietnam (n = 1). Since healthcare systems vary across geographic contexts (from country to country54,55 and from rural to urban areas56), it is possible that the findings of our reviewed papers would change if these studies were conducted in new geographic contexts. Our review indicates that future research should be including more diverse geographic contexts, such as undeveloped and developing countries.
Implications for future SC research. More qualitative and mixed-methods research is needed. A key takeaway of our review is that the majority of papers used quantitative methods (n = 14). These studies measured UX via questionnaires or through clinical vignettes. Few of the studies we reviewed investigated in depth why individuals decide to use, trust, and are satisfied with SCs. We call for more qualitative research exploring the deeper reasons behind users’ preferences. In addition, mixed methods are common in healthcare research as they can improve the research rigor and offer a more comprehensive portrayal of the studied phenomenon57,58. Though the qualitative studies and mixed-method studies in our corpus revealed how individuals’ trust in SCs, SCs’ usability, and SCs’ accuracy were measured, there were only a few such studies. To validate the qualitative data and produce more generalizable results, future research needs to complement the qualitative data with quantitative data from large samples59.
More diverse user groups should be explored. Our review suggests that future UX research should consider diverse user groups. One paper we reviewed argued that when designing new SCs, programmers should consider a diversity of user groups, as existing SCs lack support for special user groups, such as transgender individuals23. This is in line with the principles of inclusive design. Inclusive design calls for the consideration of all users’ needs without adaptation60. Prior research has highlighted the importance of inclusiveness in healthcare services61. The reviewed studies reported on some user characteristics, such as age, gender, education level, health literacy, and household income. However, the number of studies we captured is quite limited (n = 4), and there is a lack of research on how users with different characteristics (e.g., different age, gender, and health literacy levels) experience SCs. When developing healthcare apps, all health literacy levels should be thoroughly considered since individuals with different health literacy levels may have different needs62. For example, older adults may have more barriers to using healthcare technologies due to limited health and technology literacy63. The gender differences may lead to different acceptance of SCs as people of different genders may have distinct attitudes toward healthcare applications64. It is thus important to uncover the diverse needs and experiences of users of different age, gender, and health literacy levels. In the future, researchers might look into how special user groups experience SCs.
More focus on safety and ethics is needed. In healthcare, safety and security problems need to be prioritized since user information is sensitive65. Recent research on the reliability and accountability of health applications has pointed out that there are no clear measures to assess and certify SCs and guarantee their reliability and quality66, which may lead to medical errors and cause unintended harm to users due to the nature of smartphones67. Several papers in our review also highlighted potential risks of using SCs, including poor specificity22 and data security issues35,44. However, there were few suggestions regarding how to ensure the safety of current SCs. To deliver safe healthcare services, it is also necessary to consider the ethical challenges or the moral considerations68. As noted in the included studies7,35, users were always concerned about their health data privacy. However, ethical issues regarding SC’s anonymity and data collection/storage were not considered in the included studies. Few studies in our corpus offered recommendations for solving ethical challenges and enacting new policies, such as establishing a data protection policy43 or testing SCs before release7,39. It is still unclear whether the developers of SC apps have sufficient credentials, disclosed enough information to ensure transparency, and guaranteed data security. Furthermore, previous research has pointed out that AI systems need regulation and strict testing before being released to the public69 since AI technologies may introduce risks due to their opaque algorithms70. Although eight papers in the corpus described their focal SCs as AI-driven, few put forward policies to ensure users’ safety. Thus, future research should explore the kinds of policies that should be made to govern SCs, addressing SCs’ accountability, data privacy and security.
Limitations of the review process. The present paper has several methodological shortcomings. First, papers not written in English were excluded. Second, due to resource constraints, we may have missed relevant contemporary papers spurred by the fast-moving development of SCs. In fact, the ACM Digital Library’s search engine limits database users to the first 2,000 papers that appear in the search engine’s index, even if they modify their search keywords several rounds to acquire as many as possible articles as we did. To limit the risk of missing papers, we searched multiple databases as well as Google Scholar and conducted a backward–forward reference check.
Conclusion
This systematic review summarizes trends in the UX of SCs. Despite the limitations, our review contributes to the SC research by providing a thorough overview of prior UX research and discussing factors that influence SC usage. Our review presents SC users’ characteristics, individuals’ motivations to use SCs, aspects of the UX of SCs, and design implications for SCs. More user studies are required to narrow the research gaps identified in our study.
Figures & Tables
References
- 1.Millenson ML, Baldwin JL, Zipperer L, Singh H. Beyond Dr. Google: the evidence on consumer-facing digital tools for diagnosis. Diagnosis. 2018;5(3):95–105. doi: 10.1515/dx-2018-0009. [DOI] [PubMed] [Google Scholar]
- 2.Ada in the Apple App Store [Internet] Available from: https://apps.apple.com/us/app/ada-check-your-health/id1099986434.
- 3.Lupton D, Jutel A. ‘It’s like having a physician in your pocket!’A critical analysis of self-diagnosis smartphone apps. Soc Sci Med [Internet] 2015;133:128–35. doi: 10.1016/j.socscimed.2015.04.004. Available from: [DOI] [PubMed] [Google Scholar]
- 4.Fraser H, Coiera E, Wong D. Safety of patient-facing digital symptom checkers. Lancet [Internet] 2018;392(10161):2263–4. doi: 10.1016/S0140-6736(18)32819-8. Available from: [DOI] [PubMed] [Google Scholar]
- 5.Prakash AV, Das S. Intelligent Conversational Agents in Mental Healthcare Services: A Thematic Analysis of User Perceptions. Pacific Asia J Assoc Inf Syst. 2020;12(2):1. [Google Scholar]
- 6.Hassenzahl M, Tractinsky N. User experience-a research agenda. Behav Inf Technol. 2006;25(2):91–7. [Google Scholar]
- 7.Aboueid S, Meyer S, Wallace JR, Mahajan S, Chaurasia A. Young Adults’ Perspectives on the Use of Symptom Checkers for Self-Triage and Self-Diagnosis: Qualitative Study. JMIR Public Heal Surveill. 2021;7(1):e22637. doi: 10.2196/22637. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8.Mueller J, Jay C, Harper S, Davies A, Vega J, Todd C. Web use for symptom appraisal of physical health conditions: a systematic review. J Med Internet Res. 2017;19(6):e6755. doi: 10.2196/jmir.6755. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 9.Farnood A, Johnston B, Mair FS. A mixed methods systematic review of the effects of patient online self-diagnosing in the ‘smart-phone society’on the healthcare professional-patient relationship and medical authority. BMC Med Inform Decis Mak. 2020;20(1):1–14. doi: 10.1186/s12911-020-01243-6. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10.Chambers D, Cantrell AJ, Johnson M, Preston L, Baxter SK, Booth A, et al. Digital and online symptom checkers and health assessment/triage services for urgent health problems: systematic review. BMJ Open. 2019;9(8):e027743. doi: 10.1136/bmjopen-2018-027743. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 11.Gottliebsen K, Petersson G. Limited evidence of benefits of patient operated intelligent primary care triage tools: findings of a literature review. BMJ Heal care informatics. 2020;27(1) doi: 10.1136/bmjhci-2019-100114. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 12.Aboueid S, Liu RH, Desta BN, Chaurasia A. The Use of Artificially Intelligent Self-Diagnosing Digital Platforms by the General Public: Scoping Review. JMIR Med informatics. 2019;7(2):e13445. doi: 10.2196/13445. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 13.Munn Z, Peters MDJ, Stern C, Tufanaru C, McArthur A, Aromataris E. Systematic review or scoping review? Guidance for authors when choosing between a systematic or scoping review approach. BMC Med Res Methodol. 2018;18(1):1–7. doi: 10.1186/s12874-018-0611-x. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14.Powley L, McIlroy G, Simons G, Raza K. Are online symptoms checkers useful for patients with inflammatory arthritis? BMC Musculoskelet Disord. 2016;17(1):1–6. doi: 10.1186/s12891-016-1189-2. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15.Bisson LJ, Komm JT, Bernas GA, Fineberg MS, Marzo JM, Rauh MA, et al. How accurate are patients at diagnosing the cause of their knee pain with the help of a web-based symptom checker? Orthop J Sport Med. 2016;4(2):2325967116630286. doi: 10.1177/2325967116630286. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 16.Bisson LJ, Komm JT, Bernas GA, Fineberg MS, Marzo JM, Rauh MA, et al. Accuracy of a computer-based diagnostic program for ambulatory patients with knee pain. Am J Sports Med. 2014;42(10):2371–6. doi: 10.1177/0363546514541654. [DOI] [PubMed] [Google Scholar]
- 17.Hageman MGJS, Anderson J, Blok R, Bossen JKJ, Ring D. Internet self-diagnosis in hand surgery. Hand. 2015;10(3):565–9. doi: 10.1007/s11552-014-9707-x. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 18.Knitza J, Mohn J, Bergmann C, Kampylafka E, Hagen M, Bohr D, et al. Accuracy, patient-perceived usability, and acceptance of two symptom checkers (Ada and Rheport) in rheumatology: interim results from a randomized controlled crossover trial. Arthritis Res Ther. 2021;23(1):1–7. doi: 10.1186/s13075-021-02498-8. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 19.Cross S, Mourad A, Zuccon G, Koopman B. Search engines vs. symptom checkers: A comparison of their effectiveness for online health advice. Proceedings of the Web Conference 2021. 2021. pp. 206–16.
- 20.Verzantvoort NCM, Teunis T, Verheij TJM, van der Velden AW. Self-triage for acute primary care via a smartphone application: Practical, safe and efficient? PLoS One. 2018;13(6):e0199284. doi: 10.1371/journal.pone.0199284. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21.Fan X, Chao D, Zhang Z, Wang D, Li X, Tian F. Utilization of self-diagnosis health chatbots in real-world settings: case study. J Med Internet Res. 2021;23(1):e19928. doi: 10.2196/19928. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 22.Price RA, Fagbuyi D, Harris R, Hanfling D, Place F, Taylor TB, et al. Feasibility of web-based self-triage by parents of children with influenza-like illness: a cautionary tale. JAMA Pediatr. 2013;167(2):112–8. doi: 10.1001/jamapediatrics.2013.1573. [DOI] [PubMed] [Google Scholar]
- 23.You Y, Gui X. AMIA Annual Symposium Proceedings. American Medical Informatics Association; 2020. Self-Diagnosis through AI-enabled Chatbot-based Symptom Checkers: User Experiences and Design Considerations. [PMC free article] [PubMed] [Google Scholar]
- 24.Aboueid S, Meyer SB, Wallace JR, Mahajan S, Nur T, Chaurasia A. Use of symptom checkers for COVID-19-related symptoms among university students: a qualitative study. BMJ Innov. 2021;7(2) doi: 10.1136/bmjinnov-2020-000498. [DOI] [PubMed] [Google Scholar]
- 25.Marco-Ruiz L, Bønes E, de la Asunción E, Gabarron E, Aviles-Solis JC, Lee E, et al. Combining multivariate statistics and the think-aloud protocol to assess Human-Computer Interaction barriers in symptom checkers. J Biomed Inform. 2017;74:104–22. doi: 10.1016/j.jbi.2017.09.002. [DOI] [PubMed] [Google Scholar]
- 26.Ponnada S. Reimagining the COVID-19 Digital Experience: The Value of User Empowerment and Accessibility in Risk Communication. Proceedings of the 38th ACM International Conference on Design of Communication. 2020. pp. 1–3.
- 27.Schrager JD, Schuler K, Isakov AP, Wright DW, Yaffee AQ, Jacobson KL, et al. Development and usability testing of a web-based COVID-19 Self-triage platform. West J Emerg Med. 2020;21(5):1054. doi: 10.5811/westjem.2020.7.48217. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 28.Reilly N, Austin M-P. Attitudes and Engagement of Pregnant and Postnatal Women With a Web-Based Emotional Health Tool (Mummatters): Cross-sectional Study. J Med Internet Res. 2021;23(3):e18517. doi: 10.2196/18517. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 29.Nieroda ME, Lophatananon A, McMillan B, Chen L-C, Hughes J, Daniels R, et al. Online decision support tool for personalized cancer symptom checking in the community (REACT): acceptability, feasibility, and usability study. JMIR cancer. 2018;4(2):e10073. doi: 10.2196/10073. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 30.Daher K, Casas J, Khaled OA, Mugellini E. Empathic chatbot response for medical assistance. Proceedings of the 20th ACM International Conference on Intelligent Virtual Agents. 2020. pp. 1–3.
- 31.Tsai C-H, You Y, Gui X, Kou Y, Carroll JM. Exploring and Promoting Diagnostic Transparency and Explainability in Online Symptom Checkers. Proceedings of the 2021 CHI Conference on Human Factors in Computing Systems. 2021. pp. 1–17.
- 32.Li W, Fan X, Zhu H, Wu J, Teng D. Research on the Influencing Factors of User Trust Based on Artificial Intelligence Self Diagnosis System. Proceedings of the ACM Turing Celebration Conference-China. 2020. pp. 197–202.
- 33.You Y, Kou Y, Ding X, Gui X. The Medical Authority of AI: A Study of AI-enabled Consumer-Facing Health Technology. Proceedings of the 2021 CHI Conference on Human Factors in Computing Systems. 2021. pp. 1–16.
- 34.HWANG Y, KIM T, KIM J, LEE J, HONG H. Leveraging challenges of an algorithm-based symptom checker on user trust through explainable AI. 2018.
- 35.Powell J. Trust me, I’m a chatbot: how artificial intelligence in health care fails the turing test. J Med Internet Res. 2019;21(10):e16222. doi: 10.2196/16222. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 36.Miller S, Gilbert S, Virani V, Wicks P. Patients’ Utilization and Perception of an Artificial Intelligence--Based Symptom Assessment and Advice Technology in a British Primary Care Waiting Room: Exploratory Pilot Study. JMIR Hum factors. 2020;7(3):e19713. doi: 10.2196/19713. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 37.Meyer AND, Giardina TD, Spitzmueller C, Shahid U, Scott TMT, Singh H. Patient Perspectives on the Usefulness of an Artificial Intelligence–Assisted Symptom Checker: Cross-Sectional Survey Study. J Med Internet Res. 2020;22(1):e14679. doi: 10.2196/14679. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 38.Winn AN, Somai M, Fergestrom N, Crotty BH. Association of use of online symptom checkers with patients’ plans for seeking care. JAMA Netw open. 2019;2(12):e1918561–e1918561. doi: 10.1001/jamanetworkopen.2019.18561. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 39.DeForte S, Huang Y, Bourgeois T, Hussain S-A, Lin S. The association between app-administered depression assessments and suicidal ideation in user comments: retrospective observational study. JMIR mHealth uHealth. 2020;8(8):e18392. doi: 10.2196/18392. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 40.Nijland N, Cranen K, Boer H, van Gemert-Pijnen JEWC, Seydel ER. Patient use and compliance with medical advice delivered by a web-based triage system in primary care. J Telemed Telecare. 2010;16(1):8–11. doi: 10.1258/jtt.2009.001004. [DOI] [PubMed] [Google Scholar]
- 41.Luger TM, Houston TK, Suls J. Older adult experience of online diagnosis: results from a scenario-based think-aloud protocol. J Med Internet Res. 2014;16(1):e2924. doi: 10.2196/jmir.2924. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 42.Lanseng EJ, Andreassen TW. Electronic healthcare: a study of people’s readiness and attitude toward performing self‐diagnosis. Int J Serv Ind Manag. 2007.
- 43.Hua TV, Hou Y. Factors That Influence the Intention to Use Self-Diagnosis Apps in Vietnam. 2020.
- 44.Marco-Ruiz L, Bønes E, de la Asunción E, Gabarron E, Aviles-Solis JC, Lee E, et al. Combining multivariate statistics and the think-aloud protocol to assess Human-Computer Interaction barriers in symptom checkers. J Biomed Inform. 2017;74:104–22. doi: 10.1016/j.jbi.2017.09.002. [DOI] [PubMed] [Google Scholar]
- 45.Richardson WC, Berwick DM, Bisgard JC, Bristow LR, Buck CR, Cassel CK. Crossing the quality chasm: a new health system for the 21st century. Washington, DC: National Academy Press; 2001. [Google Scholar]
- 46.Cazabon D, Pande T, Sen P, Daftary A, Arsenault C, Bhatnagar H, et al. User experience and patient satisfaction with tuberculosis care in low-and middle-income countries: a systematic review. J Clin Tuberc other Mycobact Dis. 2020;19:100154. doi: 10.1016/j.jctube.2020.100154. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 47.Tolmie P, Crabtree A. Deploying research technology in the home. Proceedings of the 2008 ACM conference on Computer supported cooperative work. 2008. pp. 639–48.
- 48.Smith SP, Barefield AC. Patients meet technology: the newest in patient-centered care initiatives. Health Care Manag (Frederick) 2007;26(4):354–62. doi: 10.1097/01.HCM.0000299254.29678.50. [DOI] [PubMed] [Google Scholar]
- 49.Raja S, Hasnain M, Vadakumchery T, Hamad J, Shah R, Hoersch M. Identifying elements of patient-centered care in underserved populations: a qualitative study of patient perspectives. PLoS One. 2015;10(5):e0126708. doi: 10.1371/journal.pone.0126708. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 50.Liu B, Sundar SS. Should machines express sympathy and empathy? Experiments with a health advice chatbot. Cyberpsychology, Behav Soc Netw. 2018;21(10):625–36. doi: 10.1089/cyber.2018.0110. [DOI] [PubMed] [Google Scholar]
- 51.Li M, Mao J. Hedonic or utilitarian? Exploring the impact of communication style alignment on user’s perception of virtual health advisory services. Int J Inf Manage. 2015;35(2):229–43. [Google Scholar]
- 52.Tolmie P, Pycock J, Diggins T, MacLean A, Karsenty A. Unremarkable computing. Proceedings of the SIGCHI conference on Human factors in computing systems. 2002. pp. 399–406.
- 53.Boulos MNK. Location-based health information services: a new paradigm in personalised information delivery. Int J Health Geogr. 2003;2(1):1–11. doi: 10.1186/1476-072X-2-2. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 54.Drechsler D, Jutting J. Different countries, different needs: the role of private health insurance in developing countries. J Health Polit Policy Law. 2007;32(3):497–534. doi: 10.1215/03616878-2007-012. [DOI] [PubMed] [Google Scholar]
- 55.Reibling N, Ariaans M, Wendt C. Worlds of healthcare: a healthcare system typology of OECD countries. Health Policy (New York) 2019;123(7):611–20. doi: 10.1016/j.healthpol.2019.05.001. [DOI] [PubMed] [Google Scholar]
- 56.Qiu Y, Lu W, Guo J, Sun C, Liu X. Examining the Urban and Rural Healthcare Progress in Big Cities of China: Analysis of Monitoring Data in Dalian from 2008 to 2017. Int J Environ Res Public Health. 2020;17(4):1148. doi: 10.3390/ijerph17041148. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 57.Moffatt S, White M, Mackintosh J, Howel D. Using quantitative and qualitative data in health services research–what happens when mixed method findings conflict?[ISRCTN61522618] BMC Health Serv Res. 2006;6(1):1–10. doi: 10.1186/1472-6963-6-28. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 58.Irizarry T, Shoemake J, Nilsen ML, Czaja S, Beach S, Dabbs AD. Patient portals as a tool for health care engagement: a mixed-method study of older adults with varying levels of health literacy and prior patient portal use. J Med Internet Res. 2017;19(3):e7099. doi: 10.2196/jmir.7099. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 59.Bautista JR, Lin TTC, Theng Y-L. 2016 49th Hawaii International Conference on System Sciences (HICSS) IEEE; 2016. How and why users use social TV systems? A systematic review of user studies; pp. 3868–77. [Google Scholar]
- 60.Abascal J, Nicolle C. Moving towards inclusive design guidelines for socially and ethically aware HCI. Interact Comput. 2005;17:484–505. [Google Scholar]
- 61.Beniuk C, Ward J, Clarkson PJ. Applying inclusive design principles in the assessment of healthcare services. Proceedings of conference on design. 2011. pp. 22–36.
- 62.Kim H, Goldsmith J V, Sengupta S, Mahmood A, Powell MP, Bhatt J, et al. Mobile health application and e-health literacy: opportunities and concerns for cancer patients and caregivers. J Cancer Educ. 2019;34(1):3–8. doi: 10.1007/s13187-017-1293-5. [DOI] [PubMed] [Google Scholar]
- 63.Harris L, Thomas V, Fox M. The influence of health care literacy on the use of PHRs among older adults. Medsurg Nurs. 2015;24(4):283–6. [PubMed] [Google Scholar]
- 64.Kuo IH, Rabindran JM, Broadbent E, Lee YI, Kerse N, Stafford RMQ, et al. RO-MAN 2009-The 18th IEEE International Symposium on Robot and Human Interactive Communication. IEEE; 2009. Age and gender factors in user acceptance of healthcare robots; pp. 214–9. [Google Scholar]
- 65.Bhuyan SS, Kim H, Isehunwa OO, Kumar N, Bhatt J, Wyant DK, et al. Privacy and security issues in mobile health: Current research and future directions. Heal Policy Technol [Internet] 2017;6(2):188–91. doi: 10.1016/j.hlpt.2017.01.004. Available from: [DOI] [Google Scholar]
- 66.Grundy QH, Wang Z, Bero LA. Challenges in Assessing Mobile Health App Quality. Am J Prev Med [Internet] 2019;51(6):1051–9. doi: 10.1016/j.amepre.2016.07.009. Available from: [DOI] [PubMed] [Google Scholar]
- 67.Pine KH, Bossen C, Chen Y, Ellingsen G, Grisot M, Mazmanian M, et al. Data Work in Healthcare: Challenges for Patients, Clinicians and Administrators. Companion of the 2018 ACM Conference on Computer Supported Cooperative Work and Social Computing. 2018. pp. 433–9.
- 68.Sun TQ, Medaglia R. Mapping the challenges of Artificial Intelligence in the public sector: Evidence from public healthcare. Gov Inf Q. 2019;36(2):368–83. [Google Scholar]
- 69.Mccartney M. Margaret McCartney : AI in medicine must be rigorously tested. Bmj [Internet] 2018;361:k1752. doi: 10.1136/bmj.k1752. Available from: [DOI] [PubMed] [Google Scholar]
- 70.Macrae C. Governing the safety of artificial intelligence in healthcare. BMJ Qual Saf. 2019;28(6):495–8. doi: 10.1136/bmjqs-2019-009484. [DOI] [PubMed] [Google Scholar]