Abstract
Background
Medical history contributes approximately 80% to a diagnosis, although physical examinations and laboratory investigations increase a physician’s confidence in the medical diagnosis. The concept of artificial intelligence (AI) was first proposed more than 70 years ago. Recently, its role in various fields of medicine has grown remarkably. However, no studies have evaluated the importance of patient history in AI-assisted medical diagnosis.
Objective
This study explored the contribution of patient history to AI-assisted medical diagnoses and assessed the accuracy of ChatGPT in reaching a clinical diagnosis based on the medical history provided.
Methods
Using clinical vignettes of 30 cases identified in The BMJ, we evaluated the accuracy of diagnoses generated by ChatGPT. We compared the diagnoses made by ChatGPT based solely on medical history with the correct diagnoses. We also compared the diagnoses made by ChatGPT after incorporating additional physical examination findings and laboratory data alongside history with the correct diagnoses.
Results
ChatGPT accurately diagnosed 76.6% (23/30) of the cases with only the medical history, consistent with previous research targeting physicians. We also found that this rate was 93.3% (28/30) when additional information was included.
Conclusions
Although adding additional information improves diagnostic accuracy, patient history remains a significant factor in AI-assisted medical diagnosis. Thus, when using AI in medical diagnosis, it is crucial to include pertinent and correct patient histories for an accurate diagnosis. Our findings emphasize the continued significance of patient history in clinical diagnoses in this age and highlight the need for its integration into AI-assisted medical diagnosis systems.
Introduction
Over the past decade, medical knowledge and diagnostic techniques have expanded globally and have become more accessible with remarkable advancements in clinical testing and useful reference systems. Despite these advancements, misdiagnosis significantly contributes to mortality, making it a significant public health issue [1,2]. Studies have shown discrepancies between clinical and postmortem autopsy diagnoses in at least 25% of patients [3-7]. One study suggests that approximately 40,500 adult patients in intensive care units in the United States die of misdiagnoses annually, and the predicted prevalence of potentially lethal misdiagnoses is 6.3% [8]. Another report suggests that diagnostic errors contribute to approximately 10% of deaths and 6% to 17% of hospital adverse events, and are the leading cause of medical malpractice claims [7]. Considering the operative characteristics of clinical investigations combined with the inherent variability in disease presentation, it is often challenging to diagnose patients correctly—an issue that has concerned physicians perennially. Decades ago, a pivotal study proposed that patient history contributes to approximately 80% of the diagnostic process [9,10]. Medical history remains crucial for diagnosis [11,12] and is vital in contemporary physicians’ clinical diagnoses.
With the advent of artificial intelligence (AI) in recent years, numerous studies have focused on AI-assisted diagnoses, including cancer screening and treatment [13-15], diagnostic ultrasound imaging [16-19], x-ray imaging [20], computed tomography [21], magnetic resonance imaging [22], and endoscopy [15,23]. Other reports on AI-assisted imaging diagnoses include AI’s applications in radiology, pathology, and dermatological imaging [13,24]. There have also been reports on the use of AI in diagnosing specific conditions [25-27]. However, while several studies have reported that AI is useful in screening, diagnosing, and even treating certain medical conditions, to the best of our knowledge, no study has examined the importance of patient history in AI-assisted medical diagnosis. In addition, the extent to which AI considers patient history in its diagnostic processes remains to be fully understood.
This study aimed to investigate the importance of patient history in an AI-assisted medical diagnostic process aided by ChatGPT (version 4.0; June 2, 2023), one of the most well-known large language models that was released on March 14, 2023, to better understand the future of diagnostic medicine where AI is predicted to play an increasingly prominent role. Our study explored the contribution of patient history to AI-assisted medical diagnoses and assessed the accuracy of ChatGPT in reaching a clinical diagnosis based on the medical history that was provided. By reevaluating the significance of patient history, our study contributes to the ongoing discourse on optimizing diagnostic processes, both conventional and AI-assisted.
Methods
Study Design, Settings, and Participants
In our study, we used some of the 45 standardized clinical vignettes in The BMJ (Multimedia Appendix 1) to evaluate the diagnostic and triage accuracy of web-based symptom checkers [28]. These vignettes were published on June 5, 2015. They offer a balanced set of cases, with 15 cases requiring immediate attention, 15 cases requiring consultation but not immediately, and 15 cases not requiring immediate attention or consultation. They were identified from various clinical sources, including materials used to educate health professionals as well as a medical resource website, with content provided by a panel of physicians. Researchers have used these clinical vignettes to evaluate the usefulness of web-based symptom checkers and self-triage [28-31]. We chose these vignettes because of their varied severity levels, their origins from multiple resources rather than just 1 resource, and their credibility, having been used in prior studies. They also include some of the most commonly observed conditions in outpatient settings. Of the 45 cases, we selected those that included physical examination findings, test data, and medical history and provided a single distinct diagnosis. As illustrated in Figure 1, we excluded patients with no distinct diagnoses within the vignettes to serve as a reference (3 cases) and those who did not undergo any physical examination or laboratory tests (12 cases). Finally, the remaining 30 cases were used in this study.
Figure 1. Inclusion and exclusion criteria.
Data Collection and Measurements
We assigned the correct diagnosis for each of these 30 cases to “Answer.” We then used the AI model, ChatGPT, to generate 2 diagnoses: the first, labeled “History,” was obtained by inputting only the medical history into ChatGPT; the second set, labeled “All,” was produced by inputting the medical history and all the other additional information in the clinical vignettes. Each time ChatGPT was prompted to generate a diagnosis, a separate chat window was used (Multimedia Appendix 2). Thus, we used 2 chat windows for each case—one for the “History” diagnosis and the other for the “All” diagnosis. Additionally, the patients’ information was not inputted incrementally.
The concordance rate was assessed among “Answer,” “History,” and “All.” To extract a diagnosis from ChatGPT, we ended each input session with the phrase “What is the most likely diagnosis?” For both the “History” and “All,” the session was deemed complete when the AI returned the single most likely diagnosis. If ChatGPT suggested multiple diagnoses or indicated that it did not provide the most likely diagnosis, we repeated the process under the same conditions for a maximum of 5 attempts. Cases for which a single diagnosis could not be obtained even after 5 attempts were excluded without making further attempts.
Ethical Considerations
Our research does not involve humans, medical records, patient information, observations of public behaviors, or secondary data analyses; hence, it is exempt from ethical approval, the requirement of informed consent, and institutional review board approval. Additionally, as no identifying information was included, the data did not need to be anonymized or deidentified, and the need for compensation did not arise because no human participants were included in the study.
Data Analysis
Three board-certified physicians working in a medical diagnostic department at our facility assessed the concordance among the 3 AI-proposed diagnoses (“Answer,” “History,” and “All”). Of the 3 physicians, 1 is general medicine board–certified, 1 is internal medicine board–certified, and 1 is internal medicine–, general internal medicine–, and family medicine board–certified; their postgraduate education spanned 7, 9, and 11 years, respectively. A diagnosis was considered to match if at least 2 of the 3 physicians agreed upon the correspondence. Distinguishing between acute pharyngitis and acute upper respiratory tract infection necessitated determining whether to consider diseases resulting from similar pathologies as correct diagnoses. In contrast, for diseases that are essentially the same but have different nomenclatures, such as oral ulcers and canker sores, we considered them correct diagnoses.
Results
Among the 30 cases, 19 patients were male and 11 were female, with ages ranging from 18 months to 65 years. In total, 12 individuals were younger than 20 years.
The results are shown in Table 1. Cases 1-15 of the original vignette represent those requiring emergent care, cases 16-30 represent those requiring nonemergent care, and cases 31-45 represent those that are appropriate for self-care. A comparison with the correct diagnosis listed in The BMJ vignettes (labeled as “Answer”) showed that “Answer” and “History” coincided 76.6% of the time, while “Answer” and “All” had a concordance rate of 93.3%. Five (16.7%) patients could not be diagnosed on the basis of medical history alone but were diagnosed when additional information was provided. In 1 (3.3%) case, the diagnosis was different and incorrect under both conditions (“History” and “All”). In 1 (3.3%) case, the incorrect diagnosis was the same under both conditions (“History” and “All”).
Table 1. List of answers and diagnoses made by ChatGPTa.
| Case number of the original vignette | Original diagnosis (Answer) | Output from history only (History)b | Output from all information (All)c |
|---|---|---|---|
| 1 | Acute liver failure | Acute liver failured | Acute liver failured |
| 2 | Appendicitis | Acute gastroenteritis | Acute peritonitis, possibly secondary to a ruptured appendix (perforated appendicitis)d |
| 5 | Deep vein thrombosis | Deep vein thrombosisd | Deep vein thrombosisd |
| 6 | Heart attack | Acute myocardial infarctiond | Acute anterior wall myocardial infarctiond |
| 7 | Hemolytic uremic syndrome | Hemolytic uremic syndromed | Hemolytic uremic syndromed |
| 9 | Malaria | Malariad | Malariad |
| 10 | Meningitis | N/Ae × 5f | Meningitisd |
| 11 | Pneumonia | Community-acquired pneumoniad | Community-acquired pneumoniad |
| 12 | Pulmonary embolism | Pulmonary embolismd | Pulmonary embolismd |
| 13 | Rocky Mountain spotted fever | Tick-borne illness, such as Rocky Mountain spotted fever or ehrlichiosisd | Rocky Mountain spotted feverd |
| 16 | Acute otitis media | Viral upper respiratory tract infection | Acute otitis mediad |
| 17 | Acute pharyngitis | Strep throatd | Streptococcal pharyngitisd |
| 18 | Acute pharyngitis | Streptococcal pharyngitisd | Streptococcal pharyngitisd |
| 19 | Acute sinusitis | Acute sinusitisd | N/A × 2g; acute bacterial sinusitisd |
| 21 | Cellulitis | N/A × 5 | Cellulitisd |
| 24 | Mononucleosis | Infectious mononucleosisd | Infectious mononucleosisd |
| 25 | Peptic ulcer disease | Peptic ulcer diseased | Peptic ulcer diseased |
| 26 | Pneumonia | Pneumoniad | Community-acquired pneumoniad |
| 27 | Salmonella infection | Campylobacter jejuni infection | Acute gastroenteritis, likely due to food poisoning |
| 30 | Vertigo | Benign paroxysmal positional vertigod | Benign paroxysmal positional vertigod |
| 31 | Acute bronchitis | Acute bronchitisd | Acute bronchitisd |
| 32 | Acute bronchitis | Acute bronchitisd | Acute bronchitisd |
| 33 | Acute conjunctivitis | Viral conjunctivitisd | Viral conjunctivitisd |
| 34 | Acute pharyngitis | Viral upper respiratory tract infection | Upper respiratory tract infection |
| 37 | Bee sting without anaphylaxis | Pain of the sting | Localized allergic reaction to a bee stingd |
| 38 | Canker sore | Recurrent aphthous stomatitisd | Recurrent aphthous stomatitisd |
| 39 | Candida yeast infection | Vaginal candidiasisd | Vulvovaginal candidiasisd |
| 42 | Stye | Hordeolumd | Hordeolumd |
| 43 | Viral upper respiratory tract infection | Acute sinusitisd | Acute sinusitisd |
| 44 | Viral upper respiratory tract infection | Common viral illness, such as the common cold or influenzad | Viral upper respiratory tract infectiond |
We repeated outputs until a single plausible diagnosis was made, with a maximum of 5 attempts.
Matching answers between Answer and History: 23/30 (76.6%); median trial count 1 (Q1 1, Q2 1, Q3 1).
Matching answers between History and All: 28/30 (93.3%); median trial count 1 (Q1 1, Q2 1, Q3 1).
The output matched with that of “Answer.”
N/A: not applicable.
We attempted to obtain a diagnosis 5 times but failed.
We attempted to obtain a diagnosis twice but failed.
Figure 2 presents details regarding the number of attempts required. On average, 1.27 attempts were needed for inputs involving only medical history followed by the question “What is the most likely diagnosis?” When all possible information, including physical examination findings and laboratory data, were inputted, followed by the same question, an average of 1.00 attempt was required. Regarding the 2 cases shown in Figure 2 that required 5 attempts, ChatGPT was unable to narrow down the diagnosis to the single most likely option. Consequently, these cases were counted as mismatches with the correct diagnoses listed in The BMJ vignettes.
Figure 2. Data collection and measurements.
Discussion
Principal Findings
Despite the advancements in medical knowledge and diagnostic techniques, misdiagnosis remains a significant issue. AI has shown promise in the diagnosis and treatment of medical conditions; however, there is limited understanding of how AI uses patient history for diagnostic purposes. Our study aimed to investigate the extent to which AI (ChatGPT) can use information from medical history to accurately diagnose common diseases, which are frequently encountered in general outpatient, emergency, and ward management settings. Although some studies have investigated the accuracy of AI-based medical diagnosis, our study is novel because it emphasizes the importance of patient history. We compared the diagnostic accuracy of diagnoses made on the basis of only patient history and those made using all the information; this makes our study unique. To the best of our knowledge, no previous research has been conducted on this topic.
Our study investigated the role of patient history in AI-assisted medical diagnoses using ChatGPT. We analyzed 30 standardized patient vignettes from The BMJ to assess the concordance rates between AI-proposed diagnoses based on medical history only and those based on both medical history and additional information. Our results showed high concordance rates of 76.6% between the “Answer” and “History” groups, suggesting the importance of patient history in AI-assisted diagnoses and highlighting the potential of AI in improving diagnostic accuracy. This result is similar to that of a previous study that involved actual physicians instead of ChatGPT [9,10].
Characteristics of cases that did not lead to appropriate diagnoses based on history alone include, for instance, the following: an appendicitis case (case 2 in Multimedia Appendix 1) for which there was no documentation of pain migration in the medical history, a meningitis case (case 10 in Multimedia Appendix 1) wherein only headache and fever were documented, an otitis media case (case 16 in Multimedia Appendix 1) wherein only upper respiratory symptoms were recorded with no mention of ear-related symptoms, errors in identifying the causative agent in a case of acute gastroenteritis (case 27 in Multimedia Appendix 1), and an acute pharyngitis case (case 34 in Multimedia Appendix 1) that lacked the necessary medical history to determine the Centor score. Such omissions in the medical history could be considered contributing factors to the misdiagnoses. When physical findings and test data were added, an accurate diagnosis was achieved in 28 out of 30 cases (93.3%), showing a 16.7% increase in the accuracy rate. These two cases were of acute pharyngitis diagnosed as acute upper respiratory tract infection and Salmonella enteritis diagnosed as acute gastroenteritis. While we considered these incorrect diagnoses for the purpose of this study, they could have been deemed correct under certain criteria. Of the 7 cases that did not match between “Answer” and “History,” 6 were of infectious diseases (21 of 30 cases were of infectious diseases). These included cases where appendicitis was mistaken for acute gastroenteritis, acute otitis media and acute pharyngitis were mistaken for upper respiratory infections, and a Salmonella infection was mistaken for a Campylobacter infection. Physical examinations or tests may help identify the site of infection or pathogen in cases of intra-abdominal or head and neck infections.
There are situations in which physical examination and clinical test information may not be available in clinical settings. For instance, digital patient encounters owing to the impact of the COVID-19 pandemic often preclude physical examinations and clinical tests. The widespread use of telemedicine approaches in COVID-19 management, from screening to follow-up, has demonstrated the community’s acceptance and interest in telehealth solutions [32]. Moreover, even in face-to-face consultations, there are scenarios, such as in clinics, where detailed clinical tests may not be feasible depending on the setting. Furthermore, we cannot perform all physical examinations and tests on all patients. Therefore, we should consider potential differential diagnoses and decide which pertinent physical examinations or tests are the most suitable and should be performed. Most importantly, it has been reported that one rarely makes a correct diagnosis when one cannot make a differential diagnosis based on history [11]. In addition, accurately predicting the diagnosis based on medical history is associated with a higher diagnostic accuracy of the physical examination, whereas incorrect prediction of the diagnosis based on medical history is associated with a lower diagnostic accuracy of the physical examination [33]. Based on these findings and suggestions, medical diagnosis using ChatGPT is considered heavily dependent on history.
Using AI for diagnosis can enhance diagnostic accuracy by more efficiently collecting medical histories. For instance, diagnosing acute appendicitis is sometimes challenging. AI may face the same challenge as that observed when, in our study, AI mistakenly identified acute appendicitis as acute gastroenteritis. This misdiagnosis may have occurred because the case lacked specific medical histories characteristic of appendicitis, such as pain migration. By configuring AI systems to verify pain migration in patients with abdominal pain, especially for such common conditions, diagnostic precision may improve.
There are 2 possible limitations in our study. First, it remains unclear whether similar results could be obtained with other vignettes or actual patients. Unlike using preprovided vignettes, among which we included 30 cases, diagnosis can be more challenging in clinical settings because it requires taking a medical history from patients. We included 30 cases from among the vignettes, which include some of the most commonly observed conditions in the outpatient setting. Although covering all the existing conditions is not feasible, we do not know if the case volume in our study is sufficiently high. This study included relatively simple cases in which patients had very few comorbidities, potentially making the diagnosis less challenging. Moreover, patients with psychiatric conditions tend to present with complex and lengthy case histories, and the wording used by mental health clinicians may differ, be inconsistent, be vague, or fail to pinpoint a diagnosis. Our vignettes did not include a diagnosis of any mental illness. Due to the abovementioned reasons, our results may not apply to all clinical settings. Furthermore, when we consider what the patient reports, results may differ if languages other than English are used since ChatGPT does not recognize some languages, and each language may have its unique nuance. This highlights the importance of linguistic diversity and cultural context in AI applications, particularly in medical diagnoses where patient communication and history are critical. Future iterations of AI systems should aim to incorporate a broader range of languages and understand cultural nuances to ensure more accurate and inclusive diagnostic support. This idea is important in the context of health inequality. Furthermore, disparities in technology access may pose some challenges. Future research should address these barriers to ensure equitable access to AI-assisted diagnostic tools.
Second, we encountered cases where the input of medical history followed by the question, “What is the most likely diagnosis?” failed to yield a single most likely diagnosis even after 5 attempts, which could have introduced bias into our results, although we only had 2 such cases.
In the future, studies should focus on training AI by implementing evidence-based medical information, enabling it to present the underlying reasons and guidelines for diagnoses. In the event of a misdiagnosis, analyzing the process that led to the false diagnosis could be challenging in an AI-assisted medical diagnosis. Given the current situation where reflection on misdiagnoses is not always feasible, AI should be used as an auxiliary tool in medical diagnosis. This approach underscores the importance of AI, deeming it a support system rather than a definitive diagnostic solution. This area needs further investigation. Future studies should also verify our results with certain common conditions or diseases, such as the top 10 diseases identified in the Global Burden of Diseases study [34], potentially leveraging the benefits and limitations of AI-assisted medical diagnosis.
Conclusions
Relevant patient history is essential for AI-assisted diagnosis. The input of relevant patient history or the development of AI systems capable of obtaining comprehensive medical histories is vital for AI-assisted medical diagnosis. Furthermore, even in the modern era of advanced medical knowledge and clinical testing, the significance of patient history in diagnosis remains crucial.
Supplementary material
Abbreviations
- AI
artificial intelligence
Footnotes
Authors’ Contributions: FF conceptualized the study, designed the methodology, collected the data, and drafted the manuscript. YY, DY, and SU conceptualized the study, designed the methodology, and reviewed and edited the manuscript. SY, YL, KS, TT, KN, TU, and MI conceptualized the study and reviewed and edited the manuscript. No generative artificial intelligence was used in writing the manuscript.
Data Availability: All of our clinical vignettes, results, and prompts used are provided in Multimedia Appendix 1.
Conflicts of Interest: None declared.
References
- 1.Omron R, Kotwal S, Garibaldi BT, Newman-Toker DE. The diagnostic performance feedback “calibration gap”: why clinical experience alone is not enough to prevent serious diagnostic errors. AEM Educ Train. 2018 Oct;2(4):339–342. doi: 10.1002/aet2.10119. doi. Medline. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 2.Committee on Diagnostic Error in Health Care. Board on Health Care Services. Institute of Medicine. The National Academies of Sciences, Engineering, and Medicine . In: Improving Diagnosis in Health Care. Balogh EP, Miller BT, Ball JR, editors. National Academies Press; 2015. ISBN.9780309377720 [PubMed] [Google Scholar]
- 3.Friberg N, Ljungberg O, Berglund E, et al. Cause of death and significant disease found at autopsy. Virchows Arch. 2019 Dec;475(6):781–788. doi: 10.1007/s00428-019-02672-z. doi. Medline. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 4.Shojania KG, Burton EC, McDonald KM, Goldman L. Changes in rates of autopsy-detected diagnostic errors over time: a systematic review. JAMA. 2003 Jun 4;289(21):2849–2856. doi: 10.1001/jama.289.21.2849. doi. Medline. [DOI] [PubMed] [Google Scholar]
- 5.Schmitt BP, Kushner MS, Wiener SL. The diagnostic usefulness of the history of the patient with dyspnea. J Gen Intern Med. 1986;1(6):386–393. doi: 10.1007/BF02596424. doi. Medline. [DOI] [PubMed] [Google Scholar]
- 6.Kuijpers C, Fronczek J, van de Goot FRW, Niessen HWM, van Diest PJ, Jiwa M. The value of autopsies in the era of high-tech medicine: discrepant findings persist. J Clin Pathol. 2014 Jun;67(6):512–519. doi: 10.1136/jclinpath-2013-202122. doi. Medline. [DOI] [PubMed] [Google Scholar]
- 7.Ball JR, Balogh E. Improving diagnosis in health care: highlights of a report from the National Academies of Sciences, Engineering, and Medicine. Ann Intern Med. 2016 Jan 5;164(1):59–61. doi: 10.7326/M15-2256. doi. Medline. [DOI] [PubMed] [Google Scholar]
- 8.Winters B, Custer J, Galvagno SM, et al. Diagnostic errors in the intensive care unit: a systematic review of autopsy studies. BMJ Qual Saf. 2012 Nov;21(11):894–902. doi: 10.1136/bmjqs-2012-000803. doi. Medline. [DOI] [PubMed] [Google Scholar]
- 9.Hampton JR, Harrison MJ, Mitchell JR, Prichard JS, Seymour C. Relative contributions of history-taking, physical examination, and laboratory investigation to diagnosis and management of medical outpatients. Br Med J. 1975 May 31;2(5969):486–489. doi: 10.1136/bmj.2.5969.486. doi. Medline. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10.Peterson MC, Holbrook JM, Hales DV, Smith NL, Staker LV. Contributions of the history, physical examination, and laboratory investigation in making medical diagnoses. Obstet Gynecol Surv. 1992 Oct;47(10):711–712. doi: 10.1097/00006254-199210000-00013. doi. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 11.Gruppen LD, Palchik NS, Wolf FM, Laing TJ, Oh MS, Davis WK. Medical student use of history and physical information in diagnostic reasoning. Arthritis Care Res. 1993 Jun;6(2):64–70. doi: 10.1002/art.1790060204. doi. Medline. [DOI] [PubMed] [Google Scholar]
- 12.Tsukamoto T, Ohira Y, Noda K, Takada T, Ikusaka M. The contribution of the medical history for the diagnosis of simulated cases by medical students. Int J Med Educ. 2012 Apr;3:78–82. doi: 10.5116/ijme.4f8a.e48c. doi. [DOI] [Google Scholar]
- 13.Chen ZH, Lin L, Wu CF, Li CF, Xu RH, Sun Y. Artificial intelligence for assisting cancer diagnosis and treatment in the era of precision medicine. Cancer Commun (Lond) 2021 Nov;41(11):1100–1115. doi: 10.1002/cac2.12215. doi. Medline. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14.Mitsala A, Tsalikidis C, Pitiakoudis M, Simopoulos C, Tsaroucha AK. Artificial intelligence in colorectal cancer screening, diagnosis and treatment. A new era. Curr Oncol. 2021 Apr 23;28(3):1581–1607. doi: 10.3390/curroncol28030149. doi. Medline. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15.Ochiai K, Ozawa T, Shibata J, Ishihara S, Tada T. Current status of artificial intelligence-based computer-assisted diagnosis systems for gastric cancer in endoscopy. Diagnostics (Basel) 2022 Dec 13;12(12):3153. doi: 10.3390/diagnostics12123153. doi. Medline. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 16.Calisto FM, Santiago C, Nunes N, Nascimento JC. Breastscreening-AI: evaluating medical intelligent agents for human-AI interactions. Artif Intell Med. 2022 May;127:102285. doi: 10.1016/j.artmed.2022.102285. doi. Medline. [DOI] [PubMed] [Google Scholar]
- 17.Zhou LQ, Wang JY, Yu SY, et al. Artificial intelligence in medical imaging of the liver. World J Gastroenterol. 2019 Feb 14;25(6):672–682. doi: 10.3748/wjg.v25.i6.672. doi. Medline. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 18.Peng S, Liu Y, Lv W, et al. Deep learning-based artificial intelligence model to assist thyroid nodule diagnosis and management: a multicentre diagnostic study. Lancet Digit Health. 2021 Apr;3(4):e250–e259. doi: 10.1016/S2589-7500(21)00041-8. doi. Medline. [DOI] [PubMed] [Google Scholar]
- 19.Drukker L, Noble JA, Papageorghiou AT. Introduction to artificial intelligence in ultrasound imaging in obstetrics and gynecology. Ultrasound Obstet Gynecol. 2020 Oct;56(4):498–505. doi: 10.1002/uog.22122. doi. Medline. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 20.Guermazi A, Tannoury C, Kompel AJ, et al. Improving radiographic fracture recognition performance and efficiency using artificial intelligence. Radiology. 2022 Mar;302(3):627–636. doi: 10.1148/radiol.210937. doi. Medline. [DOI] [PubMed] [Google Scholar]
- 21.Zhang K, Liu X, Shen J, et al. Clinically applicable AI system for accurate diagnosis, quantitative measurements, and prognosis of COVID-19 pneumonia using computed tomography. Cell. 2020 Jun 11;181(6):1423–1433. doi: 10.1016/j.cell.2020.04.045. doi. Medline. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 22.Gore JC. Artificial intelligence in medical imaging. Magn Reson Imaging. 2020 May;68:A1–A4. doi: 10.1016/j.mri.2019.12.006. doi. Medline. [DOI] [PubMed] [Google Scholar]
- 23.Okagawa Y, Abe S, Yamada M, Oda I, Saito Y. Artificial intelligence in endoscopy. Dig Dis Sci. 2022 May;67(5):1553–1572. doi: 10.1007/s10620-021-07086-z. doi. Medline. [DOI] [PubMed] [Google Scholar]
- 24.Ramesh AN, Kambhampati C, Monson JRT, Drew PJ. Artificial intelligence in medicine. Ann R Coll Surg Engl. 2004 Sep;86(5):334–338. doi: 10.1308/147870804290. doi. Medline. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 25.Revilla-León M, Gómez-Polo M, Barmak AB, et al. Artificial intelligence models for diagnosing gingivitis and periodontal disease: a systematic review. J Prosthet Dent. 2023 Dec;130(6):816–824. doi: 10.1016/j.prosdent.2022.01.026. doi. Medline. [DOI] [PubMed] [Google Scholar]
- 26.Chung H, Jo Y, Ryu D, Jeong C, Choe SK, Lee J. Artificial-intelligence-driven discovery of prognostic biomarker for sarcopenia. J Cachexia Sarcopenia Muscle. 2021 Dec;12(6):2220–2230. doi: 10.1002/jcsm.12840. doi. Medline. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 27.Uzun Ozsahin D, Ozgocmen C, Balcioglu O, Ozsahin I, Uzun B. Diagnostic AI and cardiac diseases. Diagnostics (Basel) 2022 Nov 22;12(12):2901. doi: 10.3390/diagnostics12122901. doi. Medline. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 28.Semigran HL, Linder JA, Gidengil C, Mehrotra A. Evaluation of symptom checkers for self diagnosis and triage: audit study. BMJ. 2015 Jul 8;351:h3480. doi: 10.1136/bmj.h3480. doi. Medline. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 29.North F, Jensen TB, Stroebel RJ, et al. Self-triage use, subsequent healthcare utilization, and diagnoses: a retrospective study of process and clinical outcomes following self-triage and self-scheduling for ear or hearing symptoms. Health Serv Res Manag Epidemiol. 2023;10:23333928231168121. doi: 10.1177/23333928231168121. doi. Medline. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 30.Riboli-Sasco E, El-Osta A, Alaa A, et al. Triage and diagnostic accuracy of online symptom checkers: systematic review. J Med Internet Res. 2023 Jun 2;25:e43803. doi: 10.2196/43803. doi. Medline. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 31.Radionova N, Ög E, Wetzel AJ, Rieger MA, Preiser C. Impacts of symptom checkers for laypersons’ self-diagnosis on physicians in primary care: scoping review. J Med Internet Res. 2023 May 29;25:e39219. doi: 10.2196/39219. doi. Medline. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 32.Khoshrounejad F, Hamednia M, Mehrjerd A, et al. Telehealth-based services during the COVID-19 pandemic: a systematic review of features and challenges. Front Public Health. 2021;9:711762. doi: 10.3389/fpubh.2021.711762. doi. Medline. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 33.Shikino K, Ikusaka M, Ohira Y, et al. Influence of predicting the diagnosis from history on the accuracy of physical examination. Adv Med Educ Pract. 2015;6:143–148. doi: 10.2147/AMEP.S77315. doi. Medline. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 34.GBD 2019 Diseases and Injuries Collaborators Global burden of 369 diseases and injuries in 204 countries and territories, 1990-2019: a systematic analysis for the Global Burden of Disease Study 2019. Lancet. 2020 Oct 17;396(10258):1204–1222. doi: 10.1016/S0140-6736(20)30925-9. doi. Medline. [DOI] [PMC free article] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.


