Skip to main content
Frontiers in Psychology logoLink to Frontiers in Psychology
. 2024 Dec 17;15:1378904. doi: 10.3389/fpsyg.2024.1378904

Artificial intelligence in mental healthcare: transformative potential vs. the necessity of human interaction

Anithamol Babu 1,2,*, Akhil P Joseph 1
PMCID: PMC11687125  PMID: 39742049

The integration of artificial intelligence (AI) into mental healthcare represents a profound shift, merging cutting-edge technology with the intricate and deeply personal dynamics of human psychology. While AI's potential to revolutionize diagnosis, treatment, and access to mental healthcare is undeniable, it has sparked a critical debate. Can AI truly replicate the essential human touch foundational to therapeutic success, or does it enhance mental healthcare by addressing inefficiencies such as limited access, personalization, and diagnostic precision? This article argues that AI's potential in mental healthcare is transformative, complementing rather than replacing human therapists, and will outline major arguments in personalization, accessibility, ethics, and human-AI balance.

Enhancing personalization and diagnostic precision with AI

Traditional therapeutic models, while human-centered, often rely on generalizations due to limited time, resources, and the subjective nature of the diagnosis (Stein et al., 2022). Mental health conditions like depression, anxiety, and schizophrenia are typically diagnosed based on symptoms, leading to generalized treatments that often overlook individual patient needs (Wakefield, 2007). This approach risks limiting care to symptom management rather than personalized interventions. AI challenges this approach by offering unprecedented levels of personalization through data analysis (Johnson et al., 2021). AI systems can process vast data from various sources, such as speech patterns, behavioral analytics, physiological responses, and genetic information, providing a comprehensive understanding of a patient's mental health (Thakkar et al., 2024). Machine learning algorithms are capable of recognizing patterns that human therapists might overlook, offering insights into mood fluctuations, cognitive distortions, and even early signs of psychosis (Zhou et al., 2022). By analyzing data at this granular level, AI enables clinicians to tailor interventions specifically to an individual's psychological and biological makeup, thus offering highly personalized treatment plans. Recent studies have demonstrated that AI can predict patient responses to different therapeutic modalities, adjusting treatment strategies dynamically, far more accurately than traditional methods (Sezgin and McKay, 2024; Cho et al., 2020).

However, while AI offers promise for improving diagnostic precision, it has limitations, as its accuracy depends on the quality of the data it is trained on. Incomplete or biased datasets can lead to significant diagnostic errors, particularly in diverse populations, by misinterpreting symptoms or overlooking the complexity of mental health conditions. This risk is especially pronounced in mental health, where inappropriate treatments can severely impact patient wellbeing (Yan et al., 2022). For instance, a study in Ethiopia found that 39.16% of patients with severe psychiatric disorders were misdiagnosed, with rates higher among non-specialists (Ayano et al., 2021). Similarly, a Canadian study reported high misdiagnosis rates among 840 primary care patients: 65.9% for major depressive disorder, 92.7% for bipolar disorder, and over 70% for anxiety disorders (Vermani et al., 2011). Such findings underscore the inherent challenges in mental health diagnosis, which often relies on subjective doctor-patient interactions prone to inaccuracy (Yan et al., 2022). Moreover, a shortage of psychiatrists, particularly in developing countries, exacerbates the issue (Sholevar et al., 2017). In contrast, machine-based diagnoses offer several advantages, including conserving human resources, increasing efficiency, enabling large-scale assessments, and potentially reducing stigma (Ueda et al., 2024); however, over-reliance on AI without adequate human oversight risks perpetuating, rather than resolving, existing issues in mental healthcare. While AI enhances diagnostics through real-time data and predictive modeling, it must be complemented by the clinical judgment of experienced professionals, as it cannot fully capture the complexity of human emotions, behaviors, and cultural factors (Graham et al., 2019; Loscalzo et al., 2017; Khare et al., 2024). Clinicians must ensure AI remains a supportive tool, not a replacement, and address risks like biased data to safeguard patient care quality (Ueda et al., 2024).

Bridging the accessibility gap: AI as a tool for mental health equity

The issue of accessibility in mental healthcare is a pressing concern, as many individuals, particularly in underserved or rural areas, struggle to access qualified mental health professionals (Morales et al., 2020). Despite the growing awareness of mental health issues, barriers such as high costs, long wait times, and overburdened healthcare systems make therapy inaccessible for a significant portion of the population (Kourgiantakis et al., 2023). This is where AI's role as a democratizing force becomes particularly relevant. AI-driven mental health platforms, like Woebot and Wysa, offer cost-effective alternatives to traditional therapy by providing digital interventions, particularly in cognitive-behavioral therapy (Haque and Rubya, 2023). These platforms can scale therapeutic support, delivering ongoing mental healthcare to individuals who may otherwise be left without any form of assistance due to financial constraints or geographic limitations, especially where human therapists are scarce (Fitzpatrick et al., 2017).

However, the belief that AI will automatically democratize mental healthcare is overly optimistic and overlooks substantial challenges. While AI platforms can offer scalable solutions, they fail to address systemic issues related to the digital divide. Many rural and low-income populations lack the technological infrastructure—such as reliable internet access, smart devices, and digital literacy—needed to benefit from AI-driven mental health interventions (Kozelka et al., 2023). Without addressing these foundational disparities, AI cannot effectively bridge the mental healthcare gap and may, instead, deepen existing inequalities. Governments and healthcare providers must invest in AI platforms and in building the necessary infrastructure and providing digital education to ensure that the most vulnerable populations can engage with these tools. According to the World Health Organization, AI's potential to reduce disparities in mental healthcare can only be realized if these systemic barriers are addressed alongside the deployment of AI-driven solutions (Khan et al., 2023).

Precision without compromise: AI's role in predictive mental healthcare

One of the significant advantages of AI in mental healthcare is its capacity for real-time monitoring and predictive analytics, particularly in managing chronic conditions like mood disorders, and schizophrenia (Thakkar et al., 2024). AI systems can continuously track patients' behavior, mood, and cognitive patterns, identifying early warning signs of relapse or deterioration before they become noticeable to clinicians (Cho et al., 2020). This enables early intervention, which can be crucial in preventing severe crises such as suicide attempts or hospitalizations. A study by Lee et al. (2021) found that AI systems could predict mood fluctuations and relapse risk in patients with mood disorders more accurately than human clinicians. By analyzing behavioral data and patient history, AI systems can foresee when a depressive episode is likely, allowing for tailored treatments or medication adjustments that can potentially alter the course of a patient's recovery.

However, while these capabilities offer clear benefits, the psychological impact of continuous AI monitoring raises significant concerns that are often overlooked. Constant surveillance could lead to feelings of anxiety, hypervigilance, or even a loss of privacy, as patients might feel reduced to data points rather than being treated as individuals with complex emotional experiences (Joseph and Babu, 2024). This can affect the therapeutic alliance between the patient and clinician, central to effective care. If patients perceive that their every behavior is being monitored by machines, the human connection fundamental to therapy may erode, creating a sense of detachment or mistrust (Prasko et al., 2022). The ethical implications of AI-driven monitoring must be critically examined, particularly regarding how it may shift the power dynamic in therapy, with patients feeling scrutinized by technology rather than supported by a human therapist. Maintaining human oversight and ensuring that AI supports, rather than undermines, the therapeutic relationship is essential (Alowais et al., 2023).

Ethical and privacy challenges in AI-driven mental healthcare

Integrating AI into mental healthcare raises serious ethical concerns, particularly regarding data privacy and algorithmic bias, both of which pose significant challenges beyond the technological safeguards currently in place (Warrier et al., 2023). Mental health data is among the most sensitive types of personal information, and the risk of misuse or data breaches can have devastating consequences for patients, including stigmatization, loss of employment, or insurance discrimination (Seh et al., 2020). Despite strong privacy protections like encryption and anonymization, real-world cases such as the Vastaamo data breach in Finland, where 36,000 psychotherapy records were compromised, highlight the vulnerability of AI systems to exploitation (Ghanbari and Koskinen, 2024; Inkster et al., 2023). As Gentili (2021) highlights, technological interventions in complex systems like the human mind create “Bio-ethical Complexity,” “Bio-ethical Complexity” raises concerns about relying solely on AI in mental healthcare, especially as cyberattack risks grow. While robust privacy measures must be continuously updated, human error and system flaws remain significant challenges.

Algorithmic bias in AI systems necessitates thorough examination, as AI models reflect the biases present in their training data, often mirroring societal inequalities across race, gender, socioeconomic status, and culture. This bias can lead to skewed diagnoses and treatment recommendations, exacerbating healthcare disparities rather than alleviating them (Celi et al., 2022). Recent cases highlight AI bias in healthcare: a U.S. hospital algorithm assigned lower risk scores to Black patients than to white patients with similar health conditions, limiting their access to care (Ledford, 2019). Another case showed a skin cancer detection model misdiagnosing darker skin tones due to predominantly white training data, reducing accuracy for non-white patients (Krakowski et al., 2024). Such examples underscore that merely refining algorithm or incorporating diverse datasets is insufficient; systemic changes in data collection, interpretation, and application are required to capture a more comprehensive and equitable view of patient needs. AI's feedback loops can entrench biases, making them harder to eliminate over time (Ferrara, 2024). Addressing this requires not just diverse datasets but also identifying implicit biases and maintaining rigorous oversight. Without these measures, AI risks deepening, rather than reducing, healthcare disparities.

A collaborative approach: merging AI and human expertise in therapy

AI's ability to analyze extensive datasets and detect patterns that may escape human therapists offers a significant advantage, particularly in areas such as diagnostic precision and individualized care. However, AI lacks the emotional intelligence and cultural sensitivity intrinsic to human therapists, whose expertise extends beyond data to include empathy, intuition, and non-verbal communication—all critical for effective mental healthcare (Minerva and Giubilini, 2023). Excessive reliance on AI risks overshadowing the therapist's clinical judgment and intuition, potentially reducing therapy to a mechanistic process devoid of human warmth and understanding (Prasko et al., 2022). Over-dependence on AI can lead to “automation bias,” where clinicians place excessive trust in machine recommendations, which may erode their role as primary decision-makers and affect the quality of personalized care. While patient perspectives on AI are mixed—some appreciate its accessibility, while others feel it may compromise the human connection in therapy—these concerns underscore the importance of implementing patient-centered AI tools that supplement, rather than replace, therapist-patient interactions (Ali et al., 2023; Sathyan et al., 2022). Explainable AI (XAI) tools, such as SHAP (Shapley Additive exPlanations) and LIME (Local Interpretable Model-Agnostic Explanations), address these challenges by providing transparency in AI decision-making, allowing therapists to understand and validate AI insights without fully relinquishing control, while continuous professional development ensures that therapists use AI as a supportive tool rather than allowing it to dominate their decisions (Ali et al., 2023; Minerva and Giubilini, 2023). The future of AI-augmented therapy hinges on maintaining a balance between AI's precision and the therapist's empathy, fostering a collaborative model that enhances rather than diminishes the core relational elements of mental healthcare (Table 1).

Table 1.

Collaborative integration of AI and human therapists in mental healthcare.

Aspect AI's role Human therapist's role Collaborative outcome
Data analysis Processes large datasets to recognize behavioral patterns Interprets data insights with personal patient knowledge Combines AI insights with therapist understanding for a more comprehensive view of the patient's needs
Real-time monitoring Tracks mood fluctuations and behavioral changes continuously Responds empathetically to real-time patient needs Supports timely interventions, allowing the therapist to address issues as they arise
Pattern recognition Identifies patterns that may be imperceptible to humans Applies nuanced, contextual understanding Detects subtle indicators while incorporating patient history and context
Empathy and connection Lacks emotional intelligence Provides empathetic, individualized care Ensures that AI's efficiency is complemented by genuine human empathy, maintaining patient trust
AI-assisted diagnosis Offers precision in symptom analysis and diagnostic support Confirms diagnosis with holistic patient assessment Reduces diagnostic errors by merging AI precision with therapist's intuitive judgment
Patient trust Enhances treatment accuracy but may seem impersonal Builds a strong therapeutic alliance with patients Combines efficient, accurate treatment plans with a personal, trusting relationship between therapist and patient
Ethical judgment Operates within programmed constraints Makes ethical decisions based on context Aligns AI's objective analysis with human ethical judgment to provide safe and fair patient care

AI as a complementary tool in mental healthcare

AI is a transformative force in mental healthcare, but its integration must balance technological precision with human empathy, ensuring that it complements, rather than replaces, the essential therapeutic relationship. As Topol (2019) notes, the convergence of AI and human intelligence has the potential to revolutionize healthcare by harnessing both systems' strengths. Complexity Science suggests a holistic approach, integrating dimensions like ethical, philosophical, religious, cultural and emotional dimensions with technological innovations, ensuring empathy and precision coexist in mental health treatment. Addressing adoption complexities—such as regulation, scalability, cost, and practitioner acceptance—requires robust infrastructure, phased implementations, pilot programs, and AI-human collaboration models to ensure safety, privacy, and equitable access like Wysa's approach to privacy concerns and adaptability across languages and cultures (Dinesh et al., 2024). Long-term sustainability also demands updates, ethical oversight, and resources to prevent biases and inconsistent care. While AI benefits early intervention, it may affect the therapeutic alliance, with continuous monitoring risking feelings of surveillance. Thus, AI should remain a complementary tool, carefully integrated to preserve the emotional and relational elements essential to mental healthcare. This article calls for actionable steps—such as ethical AI investment and patient-centered design—to bridge human-AI gaps in mental healthcare.

Acknowledgments

We thank Kalapan Sarthy, Tata Institute of Social Sciences Guwahati Off-Campus, Jalukbari, Assam, India, for correcting and editing the language. The authors used OpenAI's ChatGPT-4 (Version 4, GPT-4 model) for assistance in language polishing.

Funding Statement

The author(s) declare that no financial support was received for the research, authorship, and/or publication of this article.

Author contributions

AB: Conceptualization, Writing – original draft, Writing – review & editing. AJ: Writing – review & editing.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher's note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

  1. Ali S., Abuhmed T., El-Sappagh S., Muhammad K., Alonso-Moral J. M., Confalonieri R., et al. (2023). Explainable Artificial Intelligence (XAI): what we know and what is left to attain Trustworthy Artificial Intelligence. Int. J. Inform. Fus. 99:101805. 10.1016/j.inffus.2023.101805 [DOI] [Google Scholar]
  2. Alowais S. A., Alghamdi S. S., Alsuhebany N., Alqahtani T., Alshaya A. I., Almohareb S. N., et al. (2023). Revolutionizing healthcare: the role of artificial intelligence in clinical practice. BMC Med. Educ. 23:689. 10.1186/s12909-023-04698-z [DOI] [PMC free article] [PubMed] [Google Scholar]
  3. Ayano G., Demelash S., Yohannes Z., Haile K., Tulu M., Assefa D., et al. (2021). Misdiagnosis, detection rate, and associated factors of severe psychiatric disorders in specialized psychiatry centers in Ethiopia. Ann. Gen. Psychiat. 20:10. 10.1186/s12991-021-00333-7 [DOI] [PMC free article] [PubMed] [Google Scholar]
  4. Celi L. A., Cellini J., Charpignon M.-L., Dee E. C., Dernoncourt F., Eber R., et al. (2022). Sources of bias in artificial intelligence that perpetuate healthcare disparities-a global review. PLoS Digit. Health 1:e0000022. 10.1371/journal.pdig.0000022 [DOI] [PMC free article] [PubMed] [Google Scholar]
  5. Cho K.-J., Kwon O., Kwon J.-M., Lee Y., Park H., Jeon K.-H., et al. (2020). Detecting patient deterioration using artificial intelligence in a rapid response system. Crit. Care Med. 48, e285–e289. 10.1097/CCM.0000000000004236 [DOI] [PubMed] [Google Scholar]
  6. Dinesh D. N., Rao M. N., Sinha C. (2024). Language adaptations of mental health interventions: user interaction comparisons with an AI-enabled conversational agent (Wysa) in English and Spanish. Digit. Health 10:20552076241255616. 10.1177/20552076241255616 [DOI] [PMC free article] [PubMed] [Google Scholar]
  7. Ferrara E. (2024). The Butterfly Effect in artificial intelligence systems: implications for AI bias and fairness. Machine Learn. Appl. 15:100525. 10.1016/j.mlwa.2024.100525 [DOI] [Google Scholar]
  8. Fitzpatrick K. K., Darcy A., Vierhile M. (2017). Delivering cognitive behavior therapy to young adults with symptoms of depression and anxiety using a fully automated conversational agent (Woebot): a randomized controlled trial. JMIR Ment. Health 4:e19. 10.2196/mental.7785 [DOI] [PMC free article] [PubMed] [Google Scholar]
  9. Gentili P. L. (2021). Why is Complexity Science valuable for reaching the goals of the UN 2030 agenda? Rendiconti Lincei. Scienze Fisiche e Naturali 32, 117–134. 10.1007/s12210-020-00972-0 [DOI] [PMC free article] [PubMed] [Google Scholar]
  10. Ghanbari H., Koskinen K. (2024). When data breach hits a psychotherapy clinic: the Vastaamo case. J. Inf. Technol. Teach. Cases 2024:20438869241258235. 10.1177/20438869241258235 [DOI] [Google Scholar]
  11. Graham S., Depp C., Lee E. E., Nebeker C., Tu X., Kim H.-C., et al. (2019). Artificial intelligence for mental health and mental illnesses: an overview. Curr. Psychiat. Rep. 21:116. 10.1007/s11920-019-1094-0 [DOI] [PMC free article] [PubMed] [Google Scholar]
  12. Haque M. D. R., Rubya S. (2023). An overview of chatbot-based mobile mental health apps: insights from app description and user reviews. JMIR mHealth uHealth 11:e44838. 10.2196/44838 [DOI] [PMC free article] [PubMed] [Google Scholar]
  13. Inkster B., Knibbs C., Bada M. (2023). Cybersecurity: a critical priority for digital mental health. Front. Digit. Health 5:1242264. 10.3389/fdgth.2023.1242264 [DOI] [PMC free article] [PubMed] [Google Scholar]
  14. Johnson K. B., Wei W.-Q., Weeraratne D., Frisse M. E., Misulis K., Rhee K., et al. (2021). Precision medicine, AI, and the future of personalized health care. Clin. Transl. Sci. 14, 86–93. 10.1111/cts.12884 [DOI] [PMC free article] [PubMed] [Google Scholar]
  15. Joseph A. P., Babu A. (2024). The unseen dilemma of AI in mental healthcare. AI Soc. 24:9. 10.1007/s00146-024-01937-9 [DOI] [Google Scholar]
  16. Khan B., Fatima H., Qureshi A., Kumar S., Hanan A., Hussain J., et al. (2023). Drawbacks of artificial intelligence and their potential solutions in the healthcare sector. Biomed. Mater. Dev. 2, 1–8. 10.1007/s44174-023-00063-2 [DOI] [PMC free article] [PubMed] [Google Scholar]
  17. Khare S. K., Blanes-Vidal V., Nadimi E. S., Acharya U. R. (2024). Emotion recognition and artificial intelligence: a systematic review (2014–2023) and research recommendations. Inf. Fus. 102:102019. 10.1016/j.inffus.2023.102019 [DOI] [Google Scholar]
  18. Kourgiantakis T., Markoulakis R., Lee E., Hussain A., Lau C., Ashcroft R., et al. (2023). Access to mental health and addiction services for youth and their families in Ontario: perspectives of parents, youth, and service providers. Int. J. Ment. Health Syst. 17:4. 10.1186/s13033-023-00572-z [DOI] [PMC free article] [PubMed] [Google Scholar]
  19. Kozelka E. E., Acquilano S. C., Al-Abdulmunem M., Guarino S., Elwyn G., Drake R. E., et al. (2023). Documenting the digital divide: identifying barriers to digital mental health access among people with serious mental illness in community settings. SSM Ment. Health 4:100241. 10.1016/j.ssmmh.2023.100241 [DOI] [Google Scholar]
  20. Krakowski I., Kim J., Cai Z. R., Daneshjou R., Lapins J., Eriksson H., et al. (2024). Human-AI interaction in skin cancer diagnosis: a systematic review and meta-analysis. NPJ Digit. Med. 7:78. 10.1038/s41746-024-01031-w [DOI] [PMC free article] [PubMed] [Google Scholar]
  21. Ledford H. (2019). Millions of black people affected by racial bias in health-care algorithms. Nature 574, 608–609. 10.1038/d41586-019-03228-6 [DOI] [PubMed] [Google Scholar]
  22. Lee E. E., Torous J., De Choudhury M., Depp C. A., Graham S. A., Kim H.-C., et al. (2021). Artificial intelligence for mental health care: clinical applications, barriers, facilitators, and artificial wisdom. Biol. Psychiat. 6, 856–864. 10.1016/j.bpsc.2021.02.001 [DOI] [PMC free article] [PubMed] [Google Scholar]
  23. Loscalzo J., Barabasi A. L., Silverman E. K. (2017). Network Medicine: Complex Systems in Human Disease and Therapeutics. Cambridge, MA: Harvard University Press. [Google Scholar]
  24. Minerva F., Giubilini A. (2023). Is AI the future of mental healthcare? Topoi Int. Rev. Philos. 42, 1–9. 10.1007/s11245-023-09932-3 [DOI] [PMC free article] [PubMed] [Google Scholar]
  25. Morales D. A., Barksdale C. L., Beckel-Mitchener A. C. (2020). A call to action to address rural mental health disparities. J. Clin. Transl. Sci. 4, 463–467. 10.1017/cts.2020.42 [DOI] [PMC free article] [PubMed] [Google Scholar]
  26. Prasko J., Ociskova M., Vanek J., Burkauskas J., Slepecky M., Bite I., et al. (2022). Managing transference and countertransference in cognitive behavioral supervision: theoretical framework and clinical application. Psychol. Res. Behav. Manag. 15, 2129–2155. 10.2147/PRBM.S369294 [DOI] [PMC free article] [PubMed] [Google Scholar]
  27. Sathyan A., Weinberg A. I., Cohen K. (2022). Interpretable AI for bio-medical applications. Compl. Eng. Syst. 2:18. 10.20517/ces.2022.41 [DOI] [PMC free article] [PubMed] [Google Scholar]
  28. Seh A. H., Zarour M., Alenezi M., Sarkar A. K., Agrawal A., Kumar R., et al. (2020). Healthcare data breaches: insights and implications. Healthcare 8:133. 10.3390/healthcare8020133 [DOI] [PMC free article] [PubMed] [Google Scholar]
  29. Sezgin E., McKay I. (2024). Behavioral health and generative AI: a perspective on future of therapies and patient care. NPJ Ment. Health Res. 3:25. 10.1038/s44184-024-00067-w [DOI] [PMC free article] [PubMed] [Google Scholar]
  30. Sholevar F., Butryn T., Bryant L., Marchionni C. (2017). The shortage of psychiatrists and other mental health providers: causes, current state, and potential solutions. Int. J. Acad. Med. 3:5. 10.4103/IJAM.IJAM_49_17 [DOI] [Google Scholar]
  31. Stein D. J., Shoptaw S. J., Vigo D. V., Lund C., Cuijpers P., Bantjes J., et al. (2022). Psychiatric diagnosis and treatment in the 21st century: paradigm shifts versus incremental integration. World Psychiat. Off. J. World Psychiat. Assoc. 21, 393–414. 10.1002/wps.20998 [DOI] [PMC free article] [PubMed] [Google Scholar]
  32. Thakkar A., Gupta A., De Sousa A. (2024). Artificial intelligence in positive mental health: a narrative review. Front. Digit. Health 6:1280235. 10.3389/fdgth.2024.1280235 [DOI] [PMC free article] [PubMed] [Google Scholar]
  33. Topol E. J. (2019). High-performance medicine: the convergence of human and artificial intelligence. Nat. Med. 25, 44–56. 10.1038/s41591-018-0300-7 [DOI] [PubMed] [Google Scholar]
  34. Ueda D., Kakinuma T., Fujita S., Kamagata K., Fushimi Y., Ito R., et al. (2024). Fairness of artificial intelligence in healthcare: review and recommendations. Japan. J. Radiol. 42, 3–15. 10.1007/s11604-023-01474-3 [DOI] [PMC free article] [PubMed] [Google Scholar]
  35. Vermani M., Marcus M., Katzman M. A. (2011). Rates of detection of mood and anxiety disorders in primary care: a descriptive, cross-sectional study. Prim. Care Compan. CNS Disord. 13:10m01013. 10.4088/PCC.10m01013 [DOI] [PMC free article] [PubMed] [Google Scholar]
  36. Wakefield J. C. (2007). The concept of mental disorder: diagnostic implications of the harmful dysfunction analysis. World Psychiat. 6, 149–156. [PMC free article] [PubMed] [Google Scholar]
  37. Warrier U., Warrier A., Khandelwal K. (2023). Ethical considerations in the use of artificial intelligence in mental health. Egypt. J. Neurol. Psychiat. Neurosurg. 59:2. 10.1186/s41983-023-00735-2 [DOI] [Google Scholar]
  38. Yan W.-J., Ruan Q.-N., Jiang K. (2022). Challenges for artificial Intelligence in recognizing mental disorders. Diagnostics 13:2. 10.3390/diagnostics13010002 [DOI] [PMC free article] [PubMed] [Google Scholar]
  39. Zhou S., Zhao J., Zhang L. (2022). Application of artificial intelligence on psychological interventions and diagnosis: an overview. Front. Psychiat. 13:811665. 10.3389/fpsyt.2022.811665 [DOI] [PMC free article] [PubMed] [Google Scholar]

Articles from Frontiers in Psychology are provided here courtesy of Frontiers Media SA

RESOURCES