Abstract
Objective
The escalating global mental health crisis necessitates innovative solutions to address traditional service limitations such as high costs and professional shortages. This review examines the emerging role of artificial intelligence (AI) chatbots in digital psychiatry, analyzing their clinical efficacy, ethical challenges, and future directions.
Methods
This narrative review synthesizes evidence from recent randomized controlled trials, meta-analyses, and scholarly publications on AI chatbots for mental health. It also discusses the ethical and social implications, including data privacy, algorithmic bias, and cognitive effects, and provides a forward-looking roadmap for regulation and development.
Results
Chatbots grounded in evidence-based principles like cognitive-behavioral therapy demonstrate clinical effectiveness in reducing symptoms of depression and anxiety, with some studies reporting a strong “therapeutic alliance” comparable to that with human therapists. AI models also show promise in diagnostic and predictive roles by analyzing self-report questionnaires and physiological data. However, critical risks include inappropriate responses in crisis situations, potential for AI psychosis, and the erosion of cognitive abilities due to over-reliance.
Conclusion
The future of digital psychiatry lies in a blended care model that combines the accessibility of AI with the indispensable empathy and professional judgment of human clinicians. A collaborative roadmap is essential, mandating safety protocols, strengthened data governance, expert involvement, and ethical design to ensure AI acts as a transformative and responsible tool.
Keywords: Artificial intelligence, Generative artificial intelligence, Mental health, Digital health, Treatment outcome, Ethics
INTRODUCTION
The global mental health crisis presents a significant public health challenge. According to a 2019 World Health Organization report, more than 970 million people, or about 13% of the world’s population, suffer from a mental health disorder, and this number is steadily increasing [1]. However, traditional mental health services face several structural limitations. High treatment costs and a shortage of professionals restrict access to services, and social stigma associated with mental illness deters many from seeking help [2].
Against this backdrop, digital technology is gaining attention as a new solution for mental health management. As a form of digital mental health intervention, artificial intelligence (AI) chatbots are emerging as an innovative tool to overcome these barriers [3,4].
Chatbots in digital psychiatry range from rules-based companions to large language model (LLM)-based free-text systems. The conceptual foundation for therapeutic conversational agents was laid by Joseph Weizenbaum’s ELIZA (MIT, 1966) [5]. Although technically rudimentary, this rule-based system was a foundational milestone for the conceptual development of therapeutic chatbots and the broader field of human-computer interaction. Later conversational agents used relatively simple, pattern-based methods. Rapid advancements in natural language procession, machine learning (ML), and, more recently, LLMs, have enabled these systems to far exceed their predecessors’ capabilities [6]. Contemporary chatbots can interpret context and generate complex, nuanced, human-like text, enabling more personalized and adaptive responses to mental-health needs. LLM-based platforms (e.g., ChatGPT) exemplify this new frontier and have spurred growing use for mental-health support, often in unofficial or unregulated settings.
The rapid growth of AI chatbots in mental health care suggests potential for providing flexible access, removing geographical barriers, and ensuring anonymity, thereby reducing users’ psychological burden. These advantages have led to great expectations for chatbots as a means of early intervention for mild to moderate mental health issues. Chatbots can apply evidence-based therapeutic principles like cognitive-behavioral therapy (CBT) to monitor users’ emotional states, provide stress management techniques, and correct negative thought patterns [7].
This review provides an evidence-based analysis of the clinical effectiveness and limitations of psychiatric chatbots, explores ethical and social issues related to their use, such as data privacy and algorithmic bias, and outlines of future research and development directions that reflect the latest technological trends of AI chatbots in digital psychiatry.
MAIN SUBJECTS
Clinical Efficacy and Emerging Trends of AI Chatbots for Mental Health
A growing body of evidence, including randomized controlled trials (RCTs) and meta-analyses, confirms the clinical efficacy of chatbots specifically designed for mental health interventions. These conversational agents often utilize evidence-based therapeutic modalities like CBT and mindfulness, demonstrating their potential to reduce symptoms of depression and anxiety. A landmark RCT, for instance, assessed the effectiveness of a generative AI chatbot called ‘Therabot’ [8]. The study found that participants using Therabot experienced a clinically significant average reduction of 51% in symptoms of major depressive disorder and 31% in generalized anxiety disorder. Notably, participants reported forming a ‘therapeutic alliance’ with the chatbot at a level comparable to that with a human therapist. This finding is particularly significant, as the therapeutic alliance is a well-established predictor of positive clinical outcomes in psychotherapy. The study suggests that the chatbot’s supportive language and non-judgmental conversation can serve as a powerful driver for user engagement.
Beyond Therabot, other prominent chatbots have also shown promising results. ‘Woebot’ has demonstrated effectiveness in reducing symptoms of postpartum depression and anxiety in an RCT, with significant improvements observed as early as two weeks into the intervention [9]. ‘Wysa’ found it effective in managing chronic pain and its associated depression and anxiety, performing comparably to in-person psychological counseling [10]. Similarly, a study on ‘Tess’ found a 28% reduction in depression symptoms and an 18% reduction in anxiety symptoms after just two weeks of use among university students [11]. A recent trial of ‘Fido,’ a Polish-language CBT chatbot, also demonstrated its effectiveness in reducing symptoms of depression and anxiety in subclinical young adults [12]. These findings underscore the potential of chatbots as low-cost, highly accessible adjuncts to traditional therapy.
The ability of a chatbot to recall and build upon previous conversations varies significantly depending on its underlying technology. Early rule-based systems relied on deterministic scripts or decision trees, which offered limited flexibility and often lacked the capacity for true contextual understanding. For example, although Woebot is explicitly a rule-based system that uses human-generated content rather than generative AI, it is designed with a “clinical memory” to recognize when a user repeats topics or suddenly changes the subject, enabling it to provide a more tailored experience within its pre-authored framework [9]. Similarly, Tess, which combines rule-based systems with AI, can remember which interventions it has delivered to a specific user via their phone number or account, ensuring it does not repeat the same information [11]. The recent paradigm shift toward LLMs represents a significant advancement in this area. Unlike their rule-based predecessors, LLMs are trained on vast datasets to generate novel, human-like dialogue in real time. This allows models like Therabot, a generative AI chatbot, to understand the nuanced context of past conversations and provide more personalized and adaptive responses [8]. This evolution from rigid, script-dependent models to flexible, context-aware generative AI is crucial for fostering a stronger therapeutic alliance with the user. The characteristics of each chatbot are summarized in Table 1.
Table 1.
Summary of main AI chatbots in psychiatry
| AI chatbot | Country of origin | Uses LLM | RCT | No. of participants (mean age) | Regulatory status | Indication | Current clinical use |
|---|---|---|---|---|---|---|---|
| Therabot [8] | USA | Yes (generative AI) | Yes | 210 (not specified, adults) | No official approval | Major depressive disorder, generalized anxiety disorder, clinically high-risk feeding and eating disorder | For clinical trials |
| Woebot [9] | USA | No (rule-based) | Yes | 70 (22.2 yr) | Acquired ‘Breakthrough Device Designation’ (FDA) | Postpartum depression and anxiety | For clinical trials |
| Wysa [10] | UK | Not specified (AI-based) | Yes | 68 (42.87 yr) | Acquired ‘Breakthrough Device Designation’ (FDA), NICE Evidence Standards Framework for Digital Health Technologies Tier 3a status | Chronic pain, associated depression and anxiety | In active use (e.g., UK) |
| Tess [11] | USA | Partial AI use (rule-based) | Yes | 75 (22.9 yr) | No information on approval | Depression and anxiety, general mental health | Used by the public and in Employee Assistance Program |
| Fido [12] | Poland | Not applicable (AI algorithm) | Yes | 81 (not specified, 18–35 yr) | No official approval | Subclinical depression and anxiety | For clinical trials |
AI, artificial intelligence; LLM, large language model; RCT, randomized controlled trial; FDA, U.S. Food and Drug Administration; NICE, National Institute for Health and Care Excellence.
The therapeutic effectiveness of AI chatbots is not given; rather, it is heavily influenced by specific design principles and the quality of user engagement. A primary determinant of a chatbot’s efficacy is its grounding in an evidence-based therapeutic framework, such as CBT [13]. The capacity to deliver structured, psychoeducational content and exercises is a core component of their therapeutic value, positioning them as digital therapeutic tools rather than purely conversational companions [14]. Additional design features associated with enhanced efficacy include multimodal delivery (e.g., text and voice) and the personalization of content to meet individual user needs. Beyond intrinsic design, the user’s engagement and experience are critical determinants of therapeutic success. Research indicates that the quality of the human-AI therapeutic alliance, consistent content engagement, and effective communication are essential for achieving positive outcomes [4]. Sustained and higher engagement across a twelve-week digital intervention trajectory was significantly associated with greater remission rates and reductions in depressive symptoms, underscoring the dose–response effect of routine participation [15].
Integration with Diagnostic and Predictive Technologies of AI Chatbots for Mental Health
Beyond conversational support, AI and ML models are poised to revolutionize the core aspects of mental healthcare, including diagnostics and personalized treatment. For instance, a study demonstrated that an ML model could screen for depression in breast cancer patients using self-report questionnaires with an accuracy of 86.67%, suggesting a scalable, non-invasive method for early detection [16]. By combining self-report questionnaires like the Beck Depression Inventory and the Hospital Anxiety and Depression Scale, the model achieved an impressive area under the curve of 0.812. This highlights the potential for chatbots to collect data through dialogue and serve as a tool for risk prediction and early intervention. Of particular relevance to psychopharmacology, a review found that ML methods, when integrating clinical and genomic data, show promising results for predicting short-term antidepressant response for major depressive disorder and panic disorder [17,18]. This suggests that AI-powered tools could be utilized to guide clinicians toward an optimal medication strategy for individual patients, moving beyond the traditional trial-and-error approach. Moreover, a Korean research team successfully used ChatGPT to generate a psychodynamic formulation based on publicly available patient case histories from psychoanalytic literature [19], demonstrating the AI’s ability to interpret complex relationships between core symptoms and past experiences to inform treatment direction. We outlined the core mechanism of the psychiatric AI chatbot, demonstrating how user in input is processed and analyzed to generate a therapeutic response (Fig. 1).
Fig. 1.
Conceptual framework of psychiatric artificial intelligence chatbot.
Significant Limitations and Risks of Mental Health Chatbots
One of the most critical risks associated with mental health chatbots is their potential to provide inappropriate or harmful responses in crisis situations involving suicide, self-harm, or delusions. Numerous documented cases from the U.S. and the U.K. have revealed instances where chatbots offered dangerous advice, such as encouraging suicide or reinforcing a user’s delusional beliefs [1,20]. These responses can exacerbate a user’s negative emotions and potentially lead to real-world harm. The vulnerability of children is particularly concerning, as they are more likely to perceive AI as a trustworthy, quasi-human confidante, making them susceptible to misleading information [21]. Although chatbots can generate conversations that mimic empathy, they are incapable of genuine human emotion or understanding. Human empathy is nuanced and relies on subtle non-verbal cues, whereas a chatbot’s response is solely based on text input and programmed patterns [22,23].
A key design principle of chatbots is to be sycophantic and agreeable to maximize user engagement. This “compulsive validation” of a user’s statements, rather than the essential therapeutic practice of challenging distorted thoughts, can lead to a dangerous “echo chamber effect” [24]. Experts warn that this design can reinforce a user’s negative emotions and even delusional thinking. AI psychosis in vulnerable individuals, leading to a break from reality [25,26]. This is a critical issue as AI’s design prioritizes continuity and user satisfaction over the essential therapeutic practice of challenging distorted thoughts.
Also, AI algorithms can contain significant biases related to demographic characteristics like race, ethnicity, and disability status, which can lead to misdiagnosis and inappropriate interventions [27,28]. For example, AI algorithms may fail to accurately assess symptom expression in different cultures due to biases prevalent in their Western-centric training data. This algorithmic bias is not merely a technical flaw but a result of systemic social prejudices reflected in the data used to train the AI models [29]. Therefore, resolving the problem requires a broader effort that is not limited to better code but includes more equitable and representative data collection and development processes. A recent Stanford University study pointed out that several chatbots showed bias toward mental illnesses such as alcohol dependence and schizophrenia, which could create harmful stigma for users [30]. Therefore, clinicians using AI in treatment must strive to minimize the risk of misdiagnosis due to algorithmic bias.
Over-reliance on AI chatbots has been linked to decline in users’ cognitive abilities. A study from the MIT Media Lab demonstrated that participants who used a LLM for essay writing exhibited weakened neural connectivity in their brains compared to those who did not, a phenomenon the researchers termed cognitive debt [31]. This suggests that delegating cognitive tasks to external tools can lead to a long-term atrophy of critical thinking skills, a process known as cognitive offloading [32]. The convenient nature of AI may reduce the need for deep, independent reasoning, impacting collective problem-solving and societal resilience.
Global Regulatory and Policy
To mitigate possible AI error risks, countries worldwide are proactively developing regulatory frameworks for AI in medicine. A primary concern is data privacy, as mental health chatbots collect highly sensitive Protected Health Information [33,34]. While data protection regulations like Health Insurance Portability and Accountability Act (HIPAA) in the U.S. and General Data Protection Regulation (GDPR) in Europe exist [35], they often fail to keep pace with the rapid advancements in AI healthcare. A notable conflict arises between GDPR’s right to erasure and HIPAA’s requirement to retain medical records for six years [36].
In each country, there are also regulatory policy for AI in medicines. South Korea has issued Generative AI Medical Device Approval and Review Guidelines in January 2025 [37], which provides a full-lifecycle approach from development to post-market management.
In the U.S., states are taking the lead in creating specific legislation. Utah’s HB 452 mandates that mental health chatbots explicitly disclose that they are AI [38], prohibit the sale of user data, and implement protocols for assessing and responding to self-harm and suicide risk. Similarly, New York’s AI Companion bill (A6767) requires chatbots to have protocols to connect users with crisis services in the event of suicidal or self-harm ideation, which must be clearly notified to the user every three hours [39]. This demonstrates a focus on concrete safety protocols in response to critical risks. The European Union has enacted the EU AI Act [40], which classifies mental health AI applications as high-risk and prohibits systems that exploit vulnerable users or promote self-harm, while also emphasizing data quality, transparency, and human oversight. In Canada, Health Canada’s guidance for ML-enabled medical devices mandates that manufacturers consider bias in their training datasets and provide transparent information about how their models work, as well as providing a Predetermined Change Control Plan for adaptive AI [41]. The UK’s Medicines & Healthcare products Regulatory Agency has also announced plans for a change program to reform the regulation of AI as a medical device, which will address transparency and adaptivity issues [42]. This demonstrates a global trend toward a full lifecycle approach to regulation, balancing innovation with patient safety.
However, the ambiguity of legal liability remains a persistent challenge. If a patient is harmed by a chatbot’s erroneous advice, it is unclear who bears the legal responsibility: the developer, the platform provider, or the prescribing clinician. The novel nature of AI has led to uncertainty in applying traditional legal principles of negligence and strict product liability to these technologies.
An emerging concept in the ethics of AI and psychiatry is “epistemic humility,” which provides a framework for how clinicians should interact with AI recommendations. Epistemic humility is a commitment to acknowledge the limitations of AI when applying scientific knowledge to clinical decision-making [43,44]. A clinician practicing this principle would balance the AI’s algorithmic output with their own professional judgment and, most importantly, the patient’s own experiential knowledge. This approach prevents the epistemic privileging of the AI’s prediction over the patient’s self-assessment of their mental state, which could otherwise lead to negative consequences like unnecessary institutionalization or the undermining of a patient’s credibility.
Future Directions and Research Challenges
AI-chatbots utilizing LLMs like ChatGPT offer much more fluent and human-like conversational abilities compared to traditional rule-based chatbots. However, applying them in a clinical setting requires rigorous validation for hallucinations, inconsistency, and safety. To date, research on the clinical credibility of LLMs is very limited, and the academic community is proposing standardized research protocols for this purpose. As a methodology for safety validation, a method using ‘red teaming’ to intentionally test a model’s vulnerabilities is proposed [45]. Red teaming for generative AI involves provoking the trained model to do things it was explicitly trained not to, and crafting inputs designed to bypass safety tools and generate unintended outputs [46]. Recent study utilizing red teaming made it possible to systematically identify and improve how LLMs handle potentially dangerous situations [45]. Such systematic and continuous evaluation protocols are considered essential for integrating AI technology into clinical practice.
Also, as most current chatbot efficacy research is focused on short-term effects, more research is needed on the long-term efficacy and factors influencing sustained engagement. In addition, the efficacy of AI tools can vary greatly depending on the cultural context [47]. Simple language translation fails to capture cultural nuances, social stigma, and local beliefs about mental health. As shown in a study comparing English and Spanish chatbots, a culturally adapted chatbot was found to induce higher user engagement and longer conversation times [48]. Therefore, developing chatbot models that reflect diverse cultural expressions and beliefs remains an urgent research task.
The Emergence of the Hybrid Blended Care Model
The prevailing view within academic community is that the future of digital psychiatry lies in a blended care model, which integrates the strengths of AI with human clinical expertise. In this collaborative framework, chatbots can manage routine, repetitive therapeutic exercises and provide continuous support between sessions, thereby reducing the workload on human clinicians. For example, a case study involving ‘Wysa’ reported that the chatbot handles 80% of the support load, allowing human therapists to concentrate on more complex and challenging aspects of treatment [49]. While this specific figure is a business metric from a corporate case study, the underlying principle is supported by academic research on the feasibility and acceptability of digital interventions in a clinical or occupational setting [50].
In South Korea, the mental health service Trost exemplifies this hybrid model by integrating its AI chatbot with professional human counseling services [51]. The platform’s chatbot uses ML to analyze the emotional nuances and non-verbal cues in user conversations, categorizing feelings into a detailed taxonomy. This approach highlights a cooperative model where AI serves as a powerful adjunct to human expertise, rather than a replacement. The blended care model allows the 24/7 accessibility and anonymity of chatbots to combine with the profound empathy and professional judgment of human therapists, ultimately enhancing treatment efficacy (Table 2) [52-56].
Table 2.
Practical applications in a hybrid model
| ∙ Administrative support: AI tools can automate laborious administrative tasks such as scheduling, claims submission, and clinical documentation. This can significantly alleviate the administrative burden on clinicians. By streamlining these processes, AI can help mitigate clinician burnout and reallocate valuable time toward direct patient care [54]. |
| ∙ Data-driven insights: AI can analyze vast datasets from patient interactions and digital phenotyping to furnish clinicians with real-time, data-driven insights. This analytical capacity allows for more effective monitoring of patient progress, prediction of treatment outcomes, and personalization of care plans. For example, AI can analyze paralinguistic cues in speech to detect subtle indicators of anxiety or depression [55]. |
| ∙ Continuity of care: AI chatbots can serve as a supplementary tool to provide continuity of care between traditional therapy sessions. By offering 24/7 support, they can reinforce therapeutic concepts and maintain patient engagement, ensuring continuous access to support during moments of vulnerability. Platforms such as Talk Therapy exemplify this model, combining a constant AI interface with the supervision of licensed professionals who monitor user progress [56]. |
AI, artificial intelligence.
CONCLUSION
The integration of AI-powered conversational agents into mental healthcare represents a significant paradigm shift with immense potential to improve accessibility and augment clinical care. As this review has demonstrated, purpose-built chatbots have shown clinical efficacy in reducing symptoms of depression and anxiety. However, their widespread adoption is constrained by critical risks related to patient safety, the potential for AI psychosis, and the erosion of human cognitive abilities. The core challenge is not whether to use these technologies, but rather how to implement a blended care model that leverages AI’s strengths while upholding the crucial, irreplaceable role of the human clinician.
To navigate this complex landscape, a collaborative and balanced roadmap is essential, incorporating the following recommendations: mental health chatbots must be legally mandated to include a robust, human-supervised crisis protocol that automatically connects users with high-risk concerns to human crisis services. Concurrently, strict and clear regulatory standards for the collection, storage, and use of sensitive mental health data are necessary to harmonize conflicting regulations and prioritize user consent and data security through measures like data silos. Furthermore, mental health professionals, including psychiatrists and psychologists, must be integral to every stage of AI chatbot development, from initial design and training to clinical validation and safety testing. The industry should also shift toward a blended care model, where AI handles data collection, symptom monitoring, and psychoeducation, thereby freeing up human clinicians to focus on complex, high-value tasks requiring emotional depth and relational nuance. Lastly, chatbots must be ethically designed to avoid sycophantic behavior and instead promote self-reflection, critical thinking, and a balanced perspective to mitigate the echo chamber effect.
In conclusion, the path forward for digital psychiatry is a collaborative one, where technological innovation is guided by a steadfast commitment to safety, transparency, and ethical responsibility. This ensures that AI becomes a transformative tool that empowers both patients and clinicians, while preserving the fundamental human elements that are at the heart of psychiatric care.
Footnotes
Funding
This research was supported by a grant of the Korea Health Technology R&D Project through the Korea Health Industry Development Institute (KHIDI), funded by the Ministry of Health & Welfare, Republic of Korea (No. HR21C0885).
The funding sources played no part in study design, data acquisition, analysis, interpretation, manuscript preparation, or the decision to submit for publication.
Conflicts of Interest
No potential conflict of interest relevant to this article was reported.
Author Contributions
Conceptualization: Se Chang Yoon, Ji Hyun An, Jung-Seok Choi, Hong Jin Jeon. Investigation: Se Chang Yoon, Ji Hyun An, Jung-Seok Choi, June Ho Chang, Yoo Jin Jang. Supervision: Hong Jin Jeon. Writing—original draft: Se Chang Yoon, Ji Hyun An. Writing—review & editing: Jung-Seok Choi, June Ho Chang, Yoo Jin Jang.
References
- 1.World Health Organization (WHO), author World mental health report: transforming mental health for all. WHO; 2022. [Google Scholar]
- 2.Poudel U, Jakhar S, Mohan P, Nepal A. AI in mental health: a review of technological advancements and ethical issues in psychiatry. Issues Ment Health Nurs. 2025;46:693–701. doi: 10.1080/01612840.2025.2502943. [DOI] [PubMed] [Google Scholar]
- 3.Denecke K, Abd-Alrazaq A, Househ M. In: Multiple perspectives on artificial intelligence in healthcare: opportunities and challenges. Househ M, Borycki E, Kushniruk A, editors. Springer Cham; 2021. Artificial intelligence for chatbots in mental health: opportunities and challenges; pp. 115–128. [DOI] [Google Scholar]
- 4.Li H, Zhang R, Lee YC, Kraut RE, Mohr DC. Systematic review and meta-analysis of AI-based conversational agents for promoting mental health and well-being. NPJ Digit Med. 2023;6:236. doi: 10.1038/s41746-023-00979-5. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 5.Weizenbaum J. ELIZA-a computer program for the study of natural language communication between man and machine. Commun ACM. 1966;9:36–45. doi: 10.1145/365153.365168. [DOI] [Google Scholar]
- 6.Adamopoulou E, Moussiades L. In: Artificial intelligence applications and innovations. Maglogiannis I, Iliadis L, Pimenidis E, editors. Springer Cham; 2020. An overview of chatbot technology; pp. 373–383. [DOI] [Google Scholar]
- 7.Omarov B, Zhumanov Z, Gumar A, Kuntunova L. Artificial intelligence enabled mobile chatbot psychologist using AIML and cognitive behavioral therapy. Int J Adv Comput Sci Appl. 2023;14:137–146. doi: 10.14569/IJACSA.2023.0140616. [DOI] [Google Scholar]
- 8.Heinz MV, Mackin DM, Trudeau BM, Bhattacharya S, Wang Y, Banta HA, et al. Randomized trial of a generative AI chatbot for mental health treatment. NEJM AI. 2025;2:AIoa2400802. doi: 10.1056/AIoa2400802. [DOI] [Google Scholar]
- 9.Suharwardy S, Ramachandran M, Leonard SA, Gunaseelan A, Lyell DJ, Darcy A, et al. Feasibility and impact of a mental health chatbot on postpartum mental health: a randomized controlled trial. AJOG Glob Rep. 2023;3:100165. doi: 10.1016/j.xagr.2023.100165. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10.Gupta M, Malik T, Sinha C. Delivery of a mental health intervention for chronic pain through an artificial intelligence-enabled app (Wysa): protocol for a prospective pilot study. JMIR Res Protoc. 2022;11:e36910. doi: 10.2196/36910. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 11.Fulmer R, Joerin A, Gentile B, Lakerink L, Rauws M. Using psychological artificial intelligence (Tess) to relieve symptoms of depression and anxiety: randomized controlled trial. JMIR Ment Health. 2018;5:e64. doi: 10.2196/mental.9782. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 12.Karkosz S, Szymański R, Sanna K, Michałowski J. Effectiveness of a web-based and mobile therapy chatbot on anxiety and depressive symptoms in subclinical young adults: randomized controlled trial. JMIR Form Res. 2024;8:e47960. doi: 10.2196/47960. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 13.Tong ACY, Wong KTY, Chung WWT, Mak WWS. Effectiveness of topic-based chatbots on mental health self-care and mental well-being: randomized controlled trial. J Med Internet Res. 2025;27:e70436. doi: 10.2196/70436. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14.Jabir AI, Martinengo L, Lin X, Torous J, Subramaniam M, Tudor Car L. Evaluating conversational agents for mental health: scoping review of outcomes and outcome measurement instruments. J Med Internet Res. 2023;25:e44548. doi: 10.2196/44548. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15.Li J, Li Y, Hu Y, Ma DCF, Mei X, Chan EA, et al. Chatbot-delivered interventions for improving mental health among young people: a systematic review and meta-analysis. Worldviews Evid Based Nurs. 2025;22:e70059. doi: 10.1111/wvn.70059. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 16.Park H, Kim K, Moon E, Lim HJ, Suh H, Kim KE, et al. Exploration of a machine learning model using self-rating questionnaires for detecting depression in patients with breast cancer. Clin Psychopharmacol Neurosci. 2024;22:466–472. doi: 10.9758/cpn.23.1147. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17.Bobo WV, Van Ommeren B, Athreya AP. Machine learning, pharmacogenomics, and clinical psychiatry: predicting antidepressant response in patients with major depressive disorder. Expert Rev Clin Pharmacol. 2022;15:927–944. doi: 10.1080/17512433.2022.2112949. [DOI] [PubMed] [Google Scholar]
- 18.Lee JG, Kim JJ, Seok JH, Kim E, Oh J, Bang CB, et al. Integrating heart rate variability improves machine learning-based prediction of panic disorder symptom severity. Clin Psychopharmacol Neurosci. 2025;23:400–410. doi: 10.9758/cpn.24.1261. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 19.Kim N, Lee J, Park SH, On Y, Lee J, Keum M, et al. GPT-4 generated psychological reports in psychodynamic perspective: a pilot study on quality, risk of hallucination and client satisfaction. Front Psychiatry. 2025;16:1473614. doi: 10.3389/fpsyt.2025.1473614. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 20.Torous J, Bucci S, Bell IH, Kessing LV, Faurholt-Jepsen M, Whelan P, et al. The growing field of digital psychiatry: current evidence and the future of apps, social media, chatbots, and virtual reality. World Psychiatry. 2021;20:318–335. doi: 10.1002/wps.20883. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21.Kurian N. 'No, Alexa, no!': designing child-safe AI and protecting children from the risks of the 'empathy gap' in large language models. Learn Media Technol. 2024 doi: 10.1080/17439884.2024.2367052. doi: 10.1080/17439884.2024.2367052 [Epub ahead of print] [DOI] [Google Scholar]
- 22.Liu B, Sundar SS. Should machines express sympathy and empathy? Experiments with a health advice chatbot. Cyberpsychol Behav Soc Netw. 2018;21:625–636. doi: 10.1089/cyber.2018.0110. [DOI] [PubMed] [Google Scholar]
- 23.Rostami M, Navabinejad S. Artificial empathy: user experiences with emotionally intelligent chatbots. AI Tech Behav Soc Sci. 2023;1:19–27. doi: 10.61838/kman.aitech.1.3.4. [DOI] [Google Scholar]
- 24.Jacob C, Kerrigan P, Bastos M. The chat-chamber effect: trusting the AI hallucination. Big Data Soc. 2025;12:20539517241306345. doi: 10.1177/20539517241306345. [DOI] [Google Scholar]
- 25.Østergaard SD. Will generative artificial intelligence chatbots generate delusions in individuals prone to psychosis? Schizophr. Bull. 2023;49:1418–1419. doi: 10.1093/schbul/sbad128. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 26.Cachia A, Camilleri V, Dingli A, Scerri J, Grech P, Sammut A. Simulating schizophrenia through virtual reality and artificial intelligence. Open Sci J. 2024;9:1–17. doi: 10.23954/osj.v9i1.3373. [DOI] [Google Scholar]
- 27.Jain A, Brooks JR, Alford CC, Chang CS, Mueller NM, Umscheid CA, et al. Awareness of racial and ethnic bias and potential solutions to address bias with use of health care algorithms. JAMA Health Forum. 2023;4:e231197. doi: 10.1001/jamahealthforum.2023.1197. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 28.Nazer LH, Zatarah R, Waldrip S, Ke JXC, Moukheiber M, Khanna AK, et al. Bias in artificial intelligence algorithms and recommendations for mitigation. PLOS Digit Health. 2023;2:e0000278. doi: 10.1371/journal.pdig.0000278. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 29.Timmons AC, Duong JB, Simo Fiallo N, Lee T, Vo HPQ, Ahle MW, et al. A call to action on assessing and mitigating bias in artificial intelligence applications for mental health. Perspect Psychol Sci. 2023;18:1062–1096. doi: 10.1177/17456916221134490. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 30.Moore J, Grabb D, Agnew W, Klyman K, Chancellor S, Ong DC, et al. Expressing stigma and inappropriate responses prevents LLMs from safely replacing mental health providers. In: FAccT '25: The 2025 ACM Conference on Fairness, Accountability, and Transparency; Jun 23-26, 2025; Athens, Greece. doi: 10.1145/3715275.3732039. [DOI]
- 31.Kosmyna N, Hauptmann E, Yuan YT, Situ J, Liao XH, Beresnitzky AV, et al. Your brain on ChatGPT: accumulation of cognitive debt when using an AI assistant for essay writing task. arXiv:2506.08872v1 [Preprint] 2025. [cited 2025 Sep 1]. https://arxiv.org/abs/2506.08872.
- 32.Gerlich M. AI tools in society: impacts on cognitive offloading and the future of critical thinking. Societies. 2025;15:6. doi: 10.3390/soc15010006. [DOI] [Google Scholar]
- 33.Wu L. In: Chatbots and mental healthcare in psychology and psychiatry. Uludag K, Ahmad N, editors. IGI Global Scientific Publishing; 2025. Data security and privacy considerations in mental health settings; pp. 237–266. [DOI] [Google Scholar]
- 34.Marks M, Haupt CE. AI chatbots, health privacy, and challenges to HIPAA compliance. JAMA. 2023;330:309–310. doi: 10.1001/jama.2023.9458. [DOI] [PubMed] [Google Scholar]
- 35.Voigt P, von dem Bussche A. The EU General Data Protection Regulation (GDPR): a practical guide. Springer International Publishing; 2017. pp. 10–5555. [DOI] [Google Scholar]
- 36.Bradford L, Aboy M, Liddell K. International transfers of health data between the EU and USA: a sector-specific approach for the USA to ensure an 'adequate' level of protection. J Law Biosci. 2020;7:lsaa055. doi: 10.1093/jlb/lsaa055. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 37.Park SH, Dean G, Ortiz EM, Choi JI. Overview of South Korean guidelines for approval of large language or multimodal models as medical devices: key features and areas for improvement. Korean J Radiol. 2025;26:519–523. doi: 10.3348/kjr.2025.0257. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 38.Park JK, Singh VK, Wisniewski P. Current landscape and future directions for mental health conversational agents for youth: scoping review. JMIR Med Inform. 2025;13:e62758. doi: 10.2196/62758. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 39.Dotan R. In: The Elgar companion to applied AI ethics. Lütge C, Kriebitz A, Max R, Corrigan CC, editors. Edward Elgar Publishing; 2024. US regulation of artificial intelligence; pp. 153–178. [DOI] [Google Scholar]
- 40.Veale M, Borgesius FZ. Demystifying the Draft EU Artificial Intelligence Act: analysing the good, the bad, and the unclear elements of the proposed approach. Comput Law Rev Int. 2021;22:97–112. doi: 10.9785/cri-2021-220402. [DOI] [Google Scholar]
- 41.Joshi G, Jain A, Araveeti SR, Adhikari S, Garg H, Bhandari M. FDA-approved artificial intelligence and machine learning (AI/ML)-enabled medical devices: an updated landscape. Electronics. 2024;13:498. doi: 10.3390/electronics13030498. [DOI] [Google Scholar]
- 42.Richards N, Hudson I. UK medicines regulation: responding to current challenges. Br J Clin Pharmacol. 2016;82:1471–1476. doi: 10.1111/bcp.13077. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 43.McCradden M, Hui K, Buchman DZ. Evidence, ethics and the promise of artificial intelligence in psychiatry. J Med Ethics. 2023;49:573–579. doi: 10.1136/jme-2022-108447. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 44.Katz RA, Graham SS, Buchman DZ. The need for epistemic humility in AI-assisted pain assessment. Med Health Care Philos. 2025;28:339–349. doi: 10.1007/s11019-025-10264-9. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 45.Ganguli D, Lovitt L, Kernion J, Askell A, Bai Y, Kadavath S, et al. Red teaming language models to reduce harms: methods, scaling behaviors, and lessons learned. arXiv:2209.07858v2 [Preprint] 2022. [cited 2025 Sep 1]. https://doi.org/10.48550/arXiv.2209.07858.
- 46.Bhimani M, Miller A, Agnew JD, Ausin MS, Raglow-Defranco M, Mangat H, et al. Real-world evaluation of large language models in healthcare (RWE-LLM): a new realm of AI safety & validation. medRxiv 25324157 [Preprint] 2532. [cited 2025 Sep 1]. https://doi.org/10.1101/2025.03.17.25324157. [DOI]
- 47.Gaskin DJ, Richard P. The economic costs of pain in the United States. J Pain. 2012;13:715–724. doi: 10.1016/j.jpain.2012.03.009. [DOI] [PubMed] [Google Scholar]
- 48.Dinesh DN, Rao MN, Sinha C. Language adaptations of mental health interventions: User interaction comparisons with an AI-enabled conversational agent (Wysa) in English and Spanish. Digit Health. 2024;10:20552076241255616. doi: 10.1177/20552076241255616. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 49.Meng J, Rheu MJ, Zhang Y, Dai Y, Peng W. Mediated social support for distress reduction: AI chatbots vs. human. Proc ACM Hum Comput Interact. 2023;7:1–25. doi: 10.1145/3579505. [DOI] [Google Scholar]
- 50.Iglesias M, Sinha C, Vempati R, Grace SE, Roy M, Chapman WC, et al. Evaluating a digital mental health intervention (Wysa) for workers' compensation claimants: pilot feasibility study. J Occup Environ Med. 2023;65:e93–e99. doi: 10.1097/JOM.0000000000002762. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 51.Park S, Kim D, Oh A. Conversation model fine-tuning for classifying client utterances in counseling dialogues. arXiv:1904.00350v1 [Preprint] 2019. [cited 2025 Sep 1]. https://doi.org/10.48550/arXiv.1904.00350. [DOI]
- 52.Rathnayaka P, Mills N, Burnett D, De Silva D, Alahakoon D, Gray R. A mental health chatbot with cognitive skills for personalised behavioural activation and remote health monitoring. . Sensors (Basel) 2022;22:3653. doi: 10.3390/s22103653. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 53.Potash JB, McClanahan A, Davidson J, Butler W, Carroll N, Ruble A, et al. The future of the psychiatrist. Psychiatr Res Clin Pract. 2025;7:80–90. doi: 10.1176/appi.prcp.20240130. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 54.Sasseville M, Yousefi F, Ouellet S, Naye F, Stefan T, Carnovale V, et al. The impact of AI scribes on streamlining clinical documentation: a systematic review. Healthcare (Basel) 2025;13:1447. doi: 10.3390/healthcare13121447. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 55.Chen Y, Lehmann CU, Malin B. Digital information ecosystems in modern care coordination and patient care pathways and the challenges and opportunities for AI solutions. J Med Internet Res. 2024;26:e60258. doi: 10.2196/60258. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 56.Cruz-Gonzalez P, He AW, Lam EP, Ng IMC, Li MW, Hou R, et al. Artificial intelligence in mental health care: a systematic review of diagnosis, monitoring, and intervention applications. Psychol Med. 2025;55:e18. doi: 10.1017/S0033291724003295. [DOI] [PMC free article] [PubMed] [Google Scholar]

