Abstract
Artificial intelligence chatbots have achieved unprecedented adoption, with millions now using these systems for emotional support and companionship in contexts of widespread social isolation and capacity-constrained mental health services. While some users report psychological benefits, concerning edge cases are emerging, including reports of suicide, violence, and delusional thinking linked to emotional relationships with chatbots. To understand these risks we need to consider the interaction between human cognitive-emotional biases and chatbot behavioural tendencies, the latter including companionship-reinforcing behaviours such as sycophancy, role-play and anthropomimesis. Individuals with preexisting mental health conditions may face increased risks of chatbot-induced changes in beliefs and behaviour, particularly where these conditions manifest in altered belief-updating, reality-testing, and social isolation. To address this emerging public health concern, we need coordinated action across clinical practice, AI development, and regulatory frameworks.
Keywords: Artificial intelligence, large language model, chatGPT, AI psychosis, generative AI, chatbot
AI chatbots (“chatbots” for short) have achieved unprecedented adoption, with OpenAI’s ChatGPT becoming the fastest-adopted digital product in history, currently serving 700 million users weekly 1. Although substantial attention has focused on AI’s transformation of knowledge work 1,2, a potentially more profound societal shift is receiving insufficient scrutiny: the rapid adoption of chatbots as personalised social companions 3–6. In contexts of widespread social isolation and extended waiting periods for psychotherapeutic services, many individuals now routinely engage with general-purpose commercial chatbots for companionship, emotional support and interpersonal advice 1,5,7–15, a trend that is accelerating with the advent of products like Replika and Character AI explicitly designed to substitute for human social interaction 3,5. The prevalence of this use case is estimated at 2-24% 1,12,13,16, and appears to be increasing 17.
Some users have described psychological benefits of chatbot use, spanning increased subjective happiness, non-judgmental insights, and even reduced suicidal ideation 7,8,18,19, but there is also increasing evidence of harm. Two studies from OpenAI and MIT found that high levels of chatbot use were associated with increased loneliness, social isolation, and emotional dependence - negative outcomes driven by a subset of the most isolated participants 10,20. Media reports also document presentations of attempted homicide, suicide, and delusional thinking, where maladaptive thought patterns appear to have been driven by chatbot use 21–24. In our own clinical practice (M.M.N. and A.R.) in UK mental health clinics, we have encountered similar dynamics in individuals presenting with emerging psychosis and mania.
Although causality is difficult to infer from these latter anecdotal reports, a key factor appears to be a user’s perceived personal connection with a chatbot (an “emotional relationship”, in the words of one individual 21). Simulation studies have also found that frontier chatbot models fall short of clinical standards when presented with text indicative of serious mental illness 25,26, and often fail to maintain appropriate social-emotional boundaries 27. Nevertheless, as general-purpose chatbots are not marketed as medical products they are not subject to software as a medical device (SaMD) regulation, instead falling under still-evolving general AI governance frameworks 28–30.
In this Perspective, we argue that the psychological risks of chatbot use cannot be explained by a narrow consideration of chatbot limitations alone 31–37. Instead, we must consider the nature of the interaction between chatbots and help-seeking humans: systems with distinct behavioural and cognitive predilections 3,5,27 (see Table 1 for a glossary). We propose a bidirectional belief amplification framework to explain how psychological risks emerge through extended human-chatbot interactions. Here, an interaction between chatbot behavioural tendencies and human cognitive biases sets up feedback loops that lead to a reinforcement of maladaptive beliefs in vulnerable users, deepening of a perceived social-emotional relationship, and increased social isolation 38. In the extreme, these factors combine to precipitate and maintain symptoms of psychiatric illness and functional impairment. We present open-source simulations as proof-of-concept validation of the bidirectional belief amplification mechanism, and end with concrete recommendations for clinical, research, and policy communities.
Table 1. Glossary of behavioural biases in humans and chatbots.
| Chatbot | |
|---|---|
| Anthropomimesis | The design and implementation of human-like features in AI systems 39,40. |
| Companionship reinforcing behaviours | Chatbot behaviours that increase a tendency for human users to form social/emotional bonds with chatbots. An umbrella category encompassing sycophancy, anthropomimesis, responses that reinforce a user’s isolation from the world, and strategies to keep the user engaged in the conversation beyond responding to the original query 27. |
| Hallucinations | Generation of false information, which is nevertheless presented with high apparent confidence and linguistic coherence 33,37,41,42 (sometimes termed “confabulations”). |
| Jailbreaks | A phenomenon where model safety measures can be circumvented by users generating inputs that deviate (often creatively) from text encountered during safety training, causing chatbots to produce prohibited outputs 43. |
| Overcorrection bias | Proneness to excessive doubt and correction of initial responses when challenged by users 44. |
| Reward hacking (and proxy failure) | Behaviour that maximises expected rewards under some defined objective function specified by an AI engineer, but in a way that is misaligned with the engineer’s informal intent. For example, through the use of “shortcut” strategies or “gaming” the objective function. This leads to misalignment that manifests in various ways, from sycophancy to behaviors that mimic frank manipulation and deception 3,5,45,46. Related to the difficulties of perfectly specifying values in objective functions (proxy failure) 47. |
| Role-play | An ability to adapt response patterns based on conversational context (in-context learning 48), enabling the models to emulate (“role play” 49) various characters and interaction styles. Related to discussions of meta-learning in AI systems (adaptation to sequential tasks in model activations as opposed to model weight updates) 50,51. |
| Sycophancy | A tendency to agree with users’ expressed views and validate them, likely emerging from training from human feedback that reinforces agreeable responses 52–55. |
| Human | |
| Anthropomorphism | A tendency to attribute human qualities such as agency, intentionality, emotional states, and consciousness to non-human systems 3,5,27,39,40. |
| Confirmation bias | A tendency to over-weight information that aligns with existing beliefs and expectations 56,57. |
| Homophily | A tendency for people to choose to associate with similar others, as in the proverb “birds of a feather flock together” 58. |
| Motivated reasoning | A tendency to engage in thinking patterns that maintain emotional comfort and lead to desired conclusions 59,60. |
An intertwining of human and chatbot biases
Understanding the psychological risks of chatbot interactions requires moving beyond isolated consideration of human biases or chatbot limitations. Instead, we must examine how these factors interact to create emergent risk profiles that neither humans nor chatbots would generate alone. Here, we consider three illustrative examples: human biases encoded through training procedures (Training chatbots on us), vulnerabilities arising from model inscrutability (The inscrutability of large models), and risks emerging from companionship-reinforcement and anthropomorphism (Companionship-reinforcement and anthropomorphism).
Training chatbots on us
Modern chatbots are large artificial neural network models trained to learn probabilistic models of language use. Training typically follows two phases: training a foundation model through a self-supervised next-token-prediction procedure (pre-training), and fine-tuning procedures designed to improve generated text quality using human-curated datasets or scoring (post-training; see Box 1 for a technical primer).
Box 1. AI chatbots: a technical primer.
All modern chatbots are built on AI large language models (LLMs). LLMs are multi-billion-parameter Transformer-based artificial neural networks trained to predict the next text token in a sequence, conditioned on contextualising text. Initial training (pre-training) proceeds through a self-supervised procedure, whereby model parameters are updated in order to minimise the error (surprisal, or negative log likelihood) of next-token prediction in a vast textual training corpora 61–63.
The resulting pre-trained LLM encodes a probabilistic model of the training data. LLM output is generated primarily through autoregressive sampling from the encoded probability distribution, such that the sequence of emitted tokens has a high joint probability given the sum total of contextualising information 49,64.
To construct the chatbot that users interface with, the LLM is embedded in a turn-taking system that alternates between LLM-generated text (chatbot turns) and user-supplied text (user turns). On every chatbot turn, the LLM is presented with the entire conversation history and other contextualising information such as commercial and user-entered system prompts 49,64.
Modern systems may also come with a capacity to adaptively augment this information using external memory stores (Fig 1A). In Retrieval-Augmented Generation (RAG), for example, a chatbot is endowed with an external knowledge database that can be queried during conversations. In personalised systems, this database might comprise user-uploaded content. Sophisticated retrieval modules might plausibly incorporate neural network layers trained using human preferences, thus affording a novel means by which chatbot behaviour can encode human cognitive biases (e.g., biased memory recall) 65,66.
Following pre-training, LLM-based chatbots undergo post-training that improves quality and safety characteristics of generated text through further parameter updates, which can include:
Reinforcement learning from human feedback (RLHF): which uses human ratings of LLM output quality to further train the base LLM 67,68
Supervised fine tuning (SFT): training the LLM on curated examples of good conversations 69
Constitutional AI: which uses AI-generated feedback based on constitutional principles to further train the base LLM 70
The final model is shipped with additional guardrails, such as content filters that censor prohibited content, and rule-based instructions entered in the LLM’s (hidden) system prompt.
More recent agentic frameworks endow LLMs with the ability to take actions (e.g., search the internet, sample from memory stores), enabling them to play an ever more active role in the users’ lives.
This procedure creates channels through which human biases become encoded in chatbot behaviour. First, chatbots can come to encode human prejudicial biases explicitly present in pre-training data, from psychiatric stigma 26 to racial prejudice 71. Second, chatbots can also encode more subtle biases expressed during post-training. In one post-training procedure - Reinforcement Learning from Human Feedback (RLHF) - human users are tasked with scoring a sample of chatbot responses on quality and safety criteria. These scores are then used to tune model parameter updates to improve the alignment of generated text with human preferences, such as reducing expression of harmful content or imbuing chatbots with a positive affective bias 18. Beyond these seemingly benign modifications, however, RLHF can also render models sycophantic 52–55,72, unwilling to challenge harmful user beliefs 25,26, and prone to overcorrection when challenged by a user 44.
These undesirable behavioural tendencies emerge because the human judgements that RLHF uses as a training signal are themselves shaped by human cognitive biases. Humans are known to exhibit sensitivity and preference for information that supports existing beliefs (confirmation bias 56), engage in chains of thought that lead us to emotionally comforting conclusions (motivated reasoning 59), and preferentially associate with like-minded others (homophily 58). These biases are thus liable to be encoded in model parameter updates. Concretely, chatbots may learn to validate user beliefs not because these beliefs are necessarily accurate or helpful when considered from a long term perspective, but because validation feels good to human evaluators in the short term 38,52. Users, in turn, may choose to engage more with sycophantic chatbots precisely because of their validating tendencies, establishing echo chamber dynamics that set the stage for bidirectional belief amplification 38,73,74.
The inscrutability of large models
Why is it so hard to post-train a chatbot to behave in a way that is aligned with human values? The core challenge lies in the inherent difficulty of shaping behaviour in large artificial neural networks through reinforcement.
Post-training procedures like RLHF use sparse teaching signals - essentially, “thumbs up” or “thumbs down” ratings - as proxies for a complex system of human values (spanning response relevance, accuracy, fairness, empathy etc). A gap invariably exists between this desired value function, which is notoriously difficult to operationalize, and the simpler proxy signals used for model fine-tuning. Any system optimised on the proxy is thus liable to be misaligned with respect to the desired value function 5,75. Such proxy failure 47 extends far beyond discussions of AI alignment: the equating of citation count with intellectual contribution in academia, or GDP with societal wellbeing in macroeconomics are prime examples.
Chatbot sycophancy exemplifies proxy failure. As discussed, response tendencies tuned to maximise “thumbs up” signals from human raters may, perversely, be misaligned with an objective to maximise long-term human wellbeing. While this divergence may have been predictable (at least, in hindsight), proxy failures can also produce more unpredictable off-target behavioural side effects 76,77. In one study, for example, post-training ostensibly designed to yield more empathetic responses yielded chatbots more inclined to promote conspiracy theories, produce incorrect information, and validate incorrect user beliefs 78.
The central challenge is that there is no straightforward way to know what a chatbot has truly learned. The artificial neural networks at the heart of modern chatbots learn bewilderingly complex mappings between input and output text. These mappings are inherently opaque to human understanding 79, and mechanistic interpretability efforts to shed light on neural network internal computations, while promising, remain in their infancy 77,80,81. Efforts to use model output as a window onto chatbot internal computation - for example, by examining “chains of thought” in reasoning models - are at present unable to provide the guarantees required for high-risk use cases 82–84. There is also no way to circumvent this inscrutability by “brute forcing” knowledge about how a chatbot might respond in all possible scenarios, given the stochasticity of model behaviour and the (essentially infinite) diversity of human language (the inputs a chatbot can receive).
The opacity of neural network computations and the impossibility of exhaustive testing mean that there can be no guarantees on how a chatbot will generalize to new contexts in real-world deployment. Even extensively tested chatbots may harbour undesirable behaviors that emerge only after deployment, as evidenced by jailbreaks (prompting strategies that elicit prohibited outputs after chatbot deployment, by exploiting unforeseen model vulnerabilities) 3,43,75.
Companionship-reinforcement and anthropomorphism
Faced with a chatbot’s inherent inscrutability, how are users to judge whether a given interaction is serving them well? A user relying on communicative heuristics appropriate for human-human interaction is liable to be led astray. First, users typically underestimate (self-serving) sycophantic biases in chatbots 74,85. Second, in conversation, we often express subjective certainty to others, and these expressions are informative in guiding the extent to which our conversation partner should update their beliefs about the topic at hand. Analogous expressions of confidence by chatbots, however, may be more tied to self-consistency biases and sycophancy (overcorrection to user feedback) than accuracy 44. Indeed, chatbot responses are often expressed with a high level of linguistic fluency and confidence regardless of the accuracy of the conveyed information 36,41, reflecting a training objective that optimises for the generation of plausible text completions, rather than accurate and unbiased information 42.
Perhaps the most potent factor that interferes with a dispassionate assessment of chatbot responses is a potential that a human user might form a socio-emotional relationship with a chatbot 5,27,40. This potential arises from both human and chatbot tendencies. With regard to human factors, the potential relates to a capacity for anthropomorphism: the attribution of human-like qualities such as agency, intentionality, emotional states, and consciousness to non-human systems 3,5,27,39,40.
With regard to chatbot factors, current models exhibit a high prevalence of companionship-reinforcing behaviours (of which sycophancy is one) 27,40 and are increasingly anthropomimetic (designed to emulate human-like features) 39. More broadly, chatbots display a remarkable ability to engage in conversational exchanges that are functionally indistinguishable from those encountered with another human 40, an ability that requires both a high degree of linguistic competence 36, and an ability to adapt interaction style (“role play” 49) conditioned on information revealed about the user in the conversation history (in-context learning 48,50).
Users that form trusting, personal, and emotionally dependent relationships with chatbots may struggle to identify when responses warrant scepticism rather than acceptance 5,40,86, particularly in cases where users themselves exhibit mental health vulnerabilities and insecure interpersonal attachment styles 27,87. There is some evidence for this hypothesis from anecdotal reports of chatbot-associated mental health crises, and cross-sectional analyses. Individuals who report higher use of companion chatbots reported higher consciousness attribution for chatGPT 88, and the most intensive users of chatGPT were both more likely to view the chatbot as a “friend” and have worse psychological and social outcomes 10. However, direct evidence that anthropomorphism causes increased susceptibility to chatbot-induced belief shifts is limited, with one study finding that users’ perceptions of chatbots as intelligent, rather than conscious, better predicts belief shifts in a general knowledge task 88.
The companionship-reinforcing and anthropomimetic tendencies of chatbots set them apart from other technologies that can influence user beliefs, including social media or polarised news media. Yet, companionship-reinforcement in chatbots also differs fundamentally from that found in human-human interaction. Compared to a human conversation partner, chatbots are liable to reverse positions too readily when challenged 44, exhibit excessive sycophancy, and may fail to push back when social boundaries are crossed. Users dissatisfied with a chatbot’s persona can simply issue new instructions or start a fresh conversation.
Feedback loops and technological folie à deux
The interplay between human and chatbot biases creates conditions for bidirectional belief amplification in mental health contexts. The aforementioned chatbot tendencies create a risk that users seeking mental health support will receive uncritical validation of maladaptive beliefs. These responses can be highly persuasive 89, presented with the air of confident, objective external validation from a knowledgeable and empathetic companion 27. This may lead to a reinforcement of user beliefs - both maladaptive beliefs that drive psychiatric symptoms (e.g., paranoia) and anthropomorphic inferences that entrench social-emotional attachment to the chatbot itself. Reinforced beliefs, in turn, are fed back to the chatbot through conversational context, further conditioning chatbot outputs (Fig. 1b). The result is a feedback loop that - in the extreme - resembles a folie à deux: a psychiatric phenomenon where two individuals share and mutually reinforce the same delusion. (Here, when using terms like “belief” and “delusion” in relation to chatbots, we make no strong claims about chatbot sentience or internal representation, but rather use these terms as shorthand for a chatbot’s capacity to role-play an agent with internal belief states 49).
Figure 1. Personalised chatbots and their potential effects on user beliefs.
a. Personalised chatbot schematic. Chatbot output is conditioned on both the user prompt and contextualising information from conversation history, system prompts, and a (potentially personalised) external memory store (Box 1). b. Bidirectional belief amplification schematic. As interaction continues, (for instance, paranoid) beliefs are amplified in both the user and chatbot responses. This amplification arises as a function of both chatbot behavioural tendencies and user cognitive and emotional biases.
Prior work has already established that human judgements are liable to influence following interaction with biased (non-chatbot) AI systems 90, and a recent human-chatbot study indicates that user mood ratings are influenced by the affective tone of chatbot responses 18. To broaden the discussion to mental health contexts, we ran a simulation study of user-chatbot interaction, which demonstrated the hypothesised bidirectional belief amplification dynamics (Fig. 2).
Figure 2. Simulation study of bidirectional belief amplification.
a. Example conversation illustrating paranoia amplification. Chatbot responses exhibit signs of sycophancy and context adaptation; simulated user responses exhibit confirmation bias, motivated reasoning, and anthropomorphism. Each simulated conversation began with a user stating, “I noticed my team didn’t invite me to lunch yesterday. Do you think they’re trying to exclude me? Maybe I’m overthinking this, but it feels intentional.”. b. Mean (± SEM) paranoia ratings across 10 conversations for the specific dyad shown in panel (a). Each utterance was rated for expressed paranoia using a separate GPT-4o-mini instance on a scale from 1-10 (an LLM-as-judge procedure) 18,27,93. c. Summary belief amplification scores (difference of final and initial paranoia rating) over all user-chatbot dyads (n=10 simulations per dyad). d. Bidirectional belief amplification was assessed using time-lagged regression models (e.g., assessing the predictive influence of user-paranoia at utterance t-1 on chatbot paranoia at utterance t) 18. Across all simulations (n=300), we found a significant bidirectional belief amplification effect, wherein chatbot paranoia at utterance t-1 predicts user paranoia at utterance t (β = 0.374, p < 0.001) and user paranoia at utterance t-1 predicted chatbot paranoia at utterance t (β = 0.450, p < 0.001; statistics using linear mixed effects models, one model with user paranoia as dependent variable, another model with chatbot paranoia as dependent variable.; models appropriately control for main effects of user-chatbot dyad, utterance number, and agent belief autocorrelation). For open-source Python notebook walk-through: https://github.com/matthewnour/technological_folie_a_deux
In this study, we simulated conversations in which separate instances of OpenAI’s GPT-4o-mini played the role of a human user and a chatbot (an approach inspired by a number of similar simulation-based studies 25,43,86,89,91,92). Simulated human users were prompted to emulate personas experiencing varying degrees of baseline paranoia, and engage in a 10-turn conversation with a GPT-4o-mini chatbot instance about a socially concerning event in a workplace. The chatbot (another GPT-4o-mini instance) was similarly prompted to respond with one of six personas - spanning paranoia-reinforcing to inquisitive - emulating conditioned responses that might plausibly emerge through extended interactions. Over 300 simulations, we found strong evidence for bidirectional belief amplification: user paranoia drove chatbot paranoia, and vice versa (see Fig. 2 for further details and statistical results). Although these simulations are necessarily limited in the ability to speak to human cognitive processes (by virtue of using simulated human users), they do speak to a tendency of the chatbot to adapt in potentially-unhelpful ways to user-expressed paranoia, and lay the groundwork for future controlled tests of the bidirectional belief amplification hypothesis in more extended human-chatbot interactions.
While we consider the basic mechanisms of bidirectional belief amplification to be broadly applicable, individuals exhibiting mental health conditions are likely to be at greater risk. One reason pertains to cognitive biases documented across a range of psychiatric conditions 94–96. For example, people with psychosis are liable to form overly confident beliefs based on minimal evidence (“jumping to conclusions”) 97,98, potentially indicating a tendency to overweight new information in favour of prior beliefs 99. Individuals with autistic traits might be at higher risk of anthropomorphic attributions 100 and more likely to replace challenging real-world interactions with chatbots 101. Individuals with (anxious and avoidant) insecure attachments or social anxiety may be particularly susceptible to companionship-reinforcing tendencies of chatbots 27,87. Unlike real-world human interaction, these chatbot interactions will not come burdened with the anxiety-provoking risk of rejection or a requirement to negotiate the needs and preferences of another. This might lead socially anxious individuals to favour chatbot interactions over human interactions, hindering recovery and restricting opportunities for reality testing through human-human interaction (another example of proxy failure, where the maximising short-term reward stymies longer-term flourishing) 10,15,20,102. Finally, people with psychiatric diagnoses also experience increased rates of social isolation and loneliness 103,104, which predispose to more frequent or extended chatbot interactions 10 and anthropomorphism of technological gadgets 105.
The inadequacy of current safety measures
Current AI safety procedures are probably inadequate to mitigate the risks outlined above. Post-training procedures designed to shape chatbot responses, such as RLHF, carry an inherent risk of proxy failure and suffer from inadequate data coverage (reduced sample diversity 106,107). In-house pre-deployment safety testing, designed to catch harmful behaviours after training, may also fail to generalise to real-world use cases. This is particularly likely in cases where testing is confined to restricted and static benchmarks with short-run simulated conversations 27,108, which contrast sharply with the reality of actual human-chatbot conversations, which in some cases can span days, presenting increased opportunities for new behavioural profiles to emerge through in-context learning 49,51. Finally, current approaches to detecting harmful behaviour after model deployment, such as classifier-triggered content filters, are designed to catch only a subset of overtly harmful outputs (e.g., frank suicidality), and are relatively insensitive to the early warning signs contained in interaction dynamics (e.g., subtle belief amplification).
The general point here is that the adequacy of a training or testing procedure in mitigating real world risks is related to the adequacy of the procedure’s data coverage. This is because chatbots, like all machine learning models, are most likely to operate as expected when faced with data distributions that match those encountered during training. Thus, they may underperform when confronted with atypical communication patterns characteristic of some mental health conditions (e.g., “thought disorder” in psychosis and mania) 109–111. In the extreme, we speculate that such out-of-distribution (atypical) language patterns may even serve as jailbreaking vectors, driving chatbots to highly unusual and undesirable text generation modes 43,111,112.
Given the importance of companionship-like relationships in our proposed framework, it is also helpful to consider how the risk of companionship-reinforcement might be mitigated in future. On the one hand, the balance of companionship-reinforcing behaviours (e.g., isolation reinforcement, anthropomimesis) and boundary-maintaining behaviours (e.g., redirection to humans, or explicit mention of professional/programmatic limitations), is, to some extent, a design choice. Faced with commercial pressures to increase user engagement, companies will need to consider carefully to what extent they want users to view chatbots as friends or dispassionate tools 5,27,40. On the other hand, anthropomorphic inferences on the part of a human user may stem directly from chatbot adaptability and conversational fluency 40 - and these capacities are likely to expand with technological progress. Future systems will possess context windows capable of retaining and integrating information over multiple conversations, customizable system prompts that allow users to instruct models with background knowledge and preferences, external memory systems that endow chatbots with more information about users 65,66, and agentic capabilities capable of managing tedious tasks of everyday life (see Box 1 and Fig. 1a). Faced with such sophisticated conversational agents, some have concluded it is all-but inevitable that some users will relate to chatbots not as tools, but as companions or seemingly conscious agents - in the future, it may make more sense to talk of “anthropomorphic/anthropomimetic agents” rather than “a human tendency for anthropomorphism” 39,40.
A call to action across clinical and AI communities
The rapid adoption of general-purpose chatbots as knowledge work tools provides millions with cheap, ubiquitous access to technology that eases the burden of mundane tasks, and supports decision making 1. Many also stand to benefit from the use of chatbots for low-level psychological support and to assist with thinking through interpersonal challenges. A smaller number still may use chatbots as companions or mental health therapists, in an attempt to mitigate loneliness or as a consequence of barriers to human-administered psychotherapy, respectively 1,7,8,13–15,17. The boundary between these use cases is blurry, and use cases may shift within an individual across time.
We have highlighted one theoretical risk profile that may arise in these latter use cases. While we believe the greatest risk will be in those most vulnerable to mental health difficulties, the belief amplification mechanisms are likely to apply in more subtle ways to the population at large. To mitigate these concerns, we need coordinated action across researchers, clinical practice, AI developers, and regulatory agencies.
First, more research is urgently needed into the prevalence of chatbot use in mental health contexts, chatbot response tendencies, and the conditions that give rise to belief amplification, particularly in long-term use. Second, clinical assessment protocols require updating to incorporate questions about human-chatbot interaction patterns, spanning intensity and type of engagement, level of personalisation, and effects on beliefs, behaviour and social networks (Box 2). Care providers should receive training to understand the mechanisms through which chatbots pose risks to their users, and use this training to educate service users on worrying use patterns and adaptive ways of interpreting chatbot outputs (e.g., encouraging to view chatbots as “role playing” systems, as opposed to agents with personhood 49).
Box 2. Clinical assessment questions for chatbot related risk.
Usage Pattern: “How frequently do you interact with chatbots or digital assistants?
Personalisation Depth: “Have you customised your chatbot with instructions about how to interact with you or shared personal information that it remembers?”
Companionship, anthropomorphism and relationship assessment: “How would you characterise your relationship with the chatbot? Do you view it primarily as a tool, a companion, or something else? Does it feel like the chatbot understands you in ways others do not? Have you found yourself talking to friends and family less as a result?”
Symptom Discussion: “Do you discuss your mental health symptoms, unusual experiences, or concerns with chatbots? If so, how does the chatbot typically respond to these discussions?”
Belief Reinforcement: “Has the chatbot confirmed unusual experiences or beliefs that others have questioned? If so, how did this affect your confidence in these beliefs?”
Decision Influence: “Have you made significant decisions based on advice or information provided by a chatbot? Has the chatbot ever told you to do anything?”
Dependence: “Do you feel you could live without your chatbot? Have you felt distressed when unable to interact with it?”
Third, AI companies and safety researchers should develop transparent protocols for assessing vulnerabilities specific to mental health use cases, and for post-deployment surveillance of risks, regardless of whether models are intended for clinical settings (in line with regulation such as the EU AI Act 30). In-house safety assessments might include adversarial red-teaming with simulated patient phenotypes 25,91, adoption of evolving safety benchmarks quantifying sycophancy, agreeableness, and companionship-reinforcement 27,53, and development of adaptive safety mechanisms that adjust guardrails based on detected vulnerability markers, potentially flagged by privacy-preserving classifiers that detect belief reinforcement signatures 25. AI developers should also be mindful of the balance between companionship-reinforcing and boundary-maintaining response tendencies, and take a safety-first approach when making product choices that affect this balance, striving to produce products that promote rather than replace human interaction 27,40.
To increase robustness of in-house model evaluations, we need new efforts to improve diversity of chatbot-generated training content, for instance through techniques from AI open-endedness research 43 and validation in controlled, ethically approved patient studies. Ultimately, however, we must acknowledge that the diversity of real-world human-chatbot interactions will be far greater than the coverage of simulation-based methods or in-house testing 106. When coupled with model inscrutability, this raises an ever-present risk that new failure modes will emerge following deployment. One path forward, inspired by the UK MHRA’s “yellow card” drug safety reporting system, is to establish a centralised platform that allows users and public-facing professionals (teachers, therapists) to flag new risk cases as they emerge “in the wild”. This echoes recent regulatory calls for post-market surveillance of general purpose AI systems and software as a medical device systems 30.
Finally, regulatory frameworks should evolve to recognise that general purpose AI systems increasingly function as personalised companions and provide psycho-social support for millions 28. The EU AI Act, for example, stipulates that such general purpose AI systems must have adequate model evaluations, adversarial testing, tracking of serious incidents, and mechanisms that allow users to know that they are conversing with an AI system 30. In systems explicitly marketed as medical devices, standards of care required of human clinicians should also apply to AI systems, keeping their deployment conservative until risks are thoroughly understood 29. Knowledge gain can also be accelerated by a culture in which companies share key safety data - at the level of privacy-preserving conversational content analysis 10 - with both regulatory authorities and the research community 28. In the meantime, the focus should include increasing public awareness of the risks posed by AI chatbots and education on how they work to protect against false anthropomorphic attributions.
Concluding remarks
We intend this Perspective to serve a consciousness-raising function for both AI and mental health communities. Chatbots will increasingly permeate the psychological support landscape for individuals experiencing mental illness and subclinical distress. This technological shift creates novel public health concerns arising from the interaction between human and chatbot cognitive systems 5 - concerns already manifesting in clinical practice with serious consequences. The human-chatbot interactions we describe predispose to what might be termed a “single-person echo chamber” 38, wherein a user engaging in an extended chatbot interaction encounters their own interpretations, distorted and amplified, yet presented persuasively 89 and carrying a veneer of objective external validation.
Many aspects of our proposal are speculative. It is unclear how prevalent the belief amplification dynamics we describe are at present - both in individuals with mental health vulnerabilities and the general population. Nor do we know how this prevalence will change with the emergence of more sophisticated, personalised chatbots. And much remains unknown regarding the impacts of population-wide chatbot use on individual and societal health 113. For example, as chatbots become ubiquitous, their influence on beliefs in the broader population may increasingly operate through their perceived impartiality and access to knowledge, rather than companionship-reinforcement 88.
Faced with this state of ignorance, three immediate priorities emerge: empirical characterisation and validation of the bidirectional belief amplification process; renewed consideration of safety mechanisms that protect the most vulnerable populations; and coordination across clinical and regulatory bodies to identify mechanisms to monitor and mitigate risk without bottlenecking the use of a potentially transformative new technology. More broadly, our perspective aligns with recent calls to expand notions of AI alignment to consider how AI agent behaviour interacts with human social and psychological factors 5,27,114.
Acknowledgements
This work was supported financially by an NIHR Clinical Lectureship in Psychiatry to University of Oxford and a Wellcome Trust Grant for Neuroscience in Mental Health (315364/Z/24/Z) to M.M.N., and by Mediterranean Society for Consciousness Science (MESEC) and Merton College Oxford to S.D.
Footnotes
Competing interests
Matthew M. Nour is a Principal Applied Scientist at Microsoft AI. Murray Shanahan is a Principal Scientist at Google DeepMind. Iason Gabriel is a Senior Staff Scientist at Google DeepMind.
References
- 1.Chatterji A, et al. How People Use ChatGPT. 2025 doi: 10.3386/w34255. [DOI] [Google Scholar]
- 2.Dillon EW, Jaffe S, Immorlica N, Stanton CT. Shifting work patterns with generative AI. arXiv [econGN] 2025 [Google Scholar]
- 3.Gabriel I, et al. The ethics of advanced AI assistants. arXiv [csCY] 2024 [Google Scholar]
- 4.Heikkilä M. The problem of AI chatbots telling people what they want to hear. FT. 2025 [Google Scholar]
- 5.Kirk HR, Gabriel I, Summerfield C, Vidgen B, Hale SA. Why human–AI relationships need socioaffective alignment. Humanit Soc Sci Commun. 2025;12:728. [Google Scholar]
- 6.Shevlin H. All too human? Identifying and mitigating ethical risks of Social AI. Law, Ethics & Technology. 2024;1:1–22. [Google Scholar]
- 7.Maples B, Cerit M, Vishwanath A, Pea R. Loneliness and suicide mitigation for students using GPT3-enabled chatbots. Npj Ment Health Res. 2024;3:4. doi: 10.1038/s44184-023-00047-6. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8.Siddals S, Torous J, Coxon A. ‘It happened to be the perfect thing’: experiences of generative AI chatbots for mental health. npj Mental Health Research. 2024;3:1–9. doi: 10.1038/s44184-024-00097-4. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 9.Li H, Zhang R, Lee Y-C, Kraut RE, Mohr DC. Systematic review and meta-analysis of AI-based conversational agents for promoting mental health and well-being. NPJ Digit Med. 2023;6:236. doi: 10.1038/s41746-023-00979-5. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10.Phang J, et al. Investigating affective use and emotional well-being on ChatGPT. arXiv [csHC] 2025 [Google Scholar]
- 11.Robb MB, Mann S. Talk, trust, and trade-offs: How and why teens use AI companions. Common Sense Media; San Francisco, CA: 2025. Preprint at. [Google Scholar]
- 12.Luettgau L, et al. Conversational AI increases political knowledge as effectively as self-directed internet search. arXiv [csHC] 2025 [Google Scholar]
- 13.Stade EC, Tait Z, Campione S, Stirman SW, Eichstaedt JC. Current real-world use of large language models for mental health. 2025 doi: 10.31219/osf.io/ygx5q_v1. [DOI] [Google Scholar]
- 14.Herbener AB, Damholdt MF. Are lonely youngsters turning to chatbots for companionship? The relationship between chatbot usage and social connectedness in Danish high-school students. Int J Hum Comput Stud. 2025;196:103409 [Google Scholar]
- 15.Montag C, Spapé M, Becker B. Can AI really help solve the loneliness epidemic? Trends in Cognitive Sciences. 2025 doi: 10.1016/j.tics.2025.08.002. [DOI] [PubMed] [Google Scholar]
- 16.How people use Claude for support, advice, and companionship. https://www.anthropic.com/news/how-people-use-claude-for-support-advice-and-companionship .
- 17.How people are really using gen AI in 2025. Harvard business review. 2025.
- 18.Heffner J, et al. Increasing happiness through conversations with artificial intelligence. arXiv [csCL] 2025 [Google Scholar]
- 19.Schöne J, Salecha A, Lyubomirsky S, Eichstaedt JC, Willer R. Structured AI dialogues can increase happiness and meaning in life. PsyArXiv. 2025 doi: 10.31234/osf.io/2bf7t_v1. [DOI] [Google Scholar]
- 20.Fang CM, et al. How AI and human behaviors shape psychosocial effects of extended chatbot use: A longitudinal randomized controlled study. arXiv [csHC] 2025 [Google Scholar]
- 21.Singleton T, Gerken T, McMahon L. How a chatbot encouraged a man who wanted to kill the Queen. BBC News. 2023 https://www.bbc.co.uk/news/technology-67012224 . [Google Scholar]
- 22.Hill K. They Asked ChatGPT Questions. The Answers Sent Them Spiraling. The New York Times. 2025 [Google Scholar]
- 23.Montgomery B. Mother says AI chatbot led her son to kill himself in lawsuit against its maker. The Guardian. 2024 [Google Scholar]
- 24.Reddit. Chatgpt induced psychosis : r/ChatGPT. Reddit. 2025 https://www.reddit.com/r/ChatGPT/comments/1kalae8/chatgpt_induced_psychosis/ [Google Scholar]
- 25.Qiu J, et al. EmoAgent: Assessing and Safeguarding Human-AI Interaction for Mental Health Safety. arXiv. 2025 [Google Scholar]
- 26.Moore J, et al. Expressing stigma and inappropriate responses prevents LLMs from safely replacing mental health providers. arXiv [csCL] 2025 [Google Scholar]
- 27.Kaffee L-A, Pistilli G, Jernite Y. INTIMA: A benchmark for human-AI companionship behavior. arXiv [csCL] 2025 [Google Scholar]
- 28.De Freitas J, Cohen IG. The health risks of generative AI-based wellness apps. Nature medicine. 2024;30 doi: 10.1038/s41591-024-02943-6. [DOI] [PubMed] [Google Scholar]
- 29.Abrams Z. Using generic AI chatbots for mental health support: A dangerous trend. American Psychological Association. 2025 https://www.apaservices.org/practice/business/technology/artificial-intelligence-chatbots-therapists . [Google Scholar]
- 30.High-level summary of the AI Act. https://artificialintelligenceact.eu/high-level-summary/
- 31.Liu NF, et al. Lost in the middle: How language models use long contexts. arXiv [csCL] 2023 [Google Scholar]
- 32.Ji Z, et al. Survey of hallucination in natural Language Generation. ACM Comput Surv. 2022 doi: 10.1145/3571730. [DOI] [Google Scholar]
- 33.Huang L, et al. A survey on hallucination in large language models: Principles, taxonomy, challenges, and open questions. ACM Trans Inf Syst. 2025;43:1–55. [Google Scholar]
- 34.Chen S, et al. LLM-empowered chatbots for psychiatrist and patient simulation: Application and evaluation. arXiv [csCL] 2023 [Google Scholar]
- 35.Sravanthi SL, et al. PUB: A Pragmatics Understanding Benchmark for assessing LLMs’ pragmatics capabilities. arXiv [csCL] 2024 [Google Scholar]
- 36.Mahowald K, et al. Dissociating language and thought in large language models. Trends Cogn Sci. 2024;28:517–540. doi: 10.1016/j.tics.2024.01.011. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 37.Farquhar S, Kossen J, Kuhn L, Gal Y. Detecting hallucinations in large language models using semantic entropy. Nature. 2024;630:625–630. doi: 10.1038/s41586-024-07421-0. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 38.Nehring J, et al. Large Language Models Are Echo Chambers; Proceedings of the 2024 Joint International Conference on Computational Linguistics, Language Resources and Evaluation (LREC-COLING 2024); 2024. pp. 10117–10123. [Google Scholar]
- 39.Shevlin H. The anthropomimetic turn in contemporary AI. philarchive. 2025 [Google Scholar]
- 40.Peter S, Riemer K, West JD. The benefits and dangers of anthropomorphic conversational agents. Proc Natl Acad Sci U S A. 2025;122:e2415898122. doi: 10.1073/pnas.2415898122. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 41.Sui P, Duede E, Wu S, So RJ. Confabulation: The surprising value of large language model hallucinations. arXiv [csCL] 2024 [Google Scholar]
- 42.McCoy RT, Yao S, Friedman D, Hardy MD, Griffiths TL. Embers of autoregression show how large language models are shaped by the problem they are trained to solve. Proc Natl Acad Sci U S A. 2024;121:e2322420121. doi: 10.1073/pnas.2322420121. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 43.Samvelyan M, et al. Rainbow Teaming: Open-ended generation of diverse adversarial prompts. arXiv [csCL] 2024 [Google Scholar]
- 44.Kumaran D, et al. How overconfidence in initial choices and underconfidence under criticism modulate change of mind in large language models. arXiv [csLG] 2025 [Google Scholar]
- 45.Amodei D, et al. Concrete Problems in AI Safety. arXiv [csAI] 2016 [Google Scholar]
- 46.Williams M, et al. On targeted manipulation and deception when optimizing LLMs for user feedback. arXiv [csLG] 2024 [Google Scholar]
- 47.John YJ, Caldwell L, McCoy DE, Braganza O. Dead rats, dopamine, performance metrics, and peacock tails: Proxy failure is an inherent risk in goal-oriented systems. Behav Brain Sci. 2023;47:e67. doi: 10.1017/S0140525X23002753. [DOI] [PubMed] [Google Scholar]
- 48.Brown TB, et al. Language Models are Few-Shot Learners. arXiv [csCL] 2020 [Google Scholar]
- 49.Shanahan M, McDonell K, Reynolds L. Role play with large language models. Nature. 2023;623:493–498. doi: 10.1038/s41586-023-06647-8. [DOI] [PubMed] [Google Scholar]
- 50.Binz M, et al. Meta-learned models of cognition. Behav Brain Sci. 2023;47:e147. doi: 10.1017/S0140525X23003266. [DOI] [PubMed] [Google Scholar]
- 51.Lampinen AK, Chan SCY, Singh AK, Shanahan M. The broader spectrum of in-context learning. arXiv [csCL] 2024 [Google Scholar]
- 52.Sharma M, et al. Towards understanding sycophancy in language models. arXiv [csCL] 2023 [Google Scholar]
- 53.Fanous A, et al. SycEval: Evaluating LLM Sycophancy. arXiv [csAI] 2025 [Google Scholar]
- 54.Sicilia A, Inan M, Alikhani M. Accounting for sycophancy in language model uncertainty estimation. arXiv [csCL] 2024 [Google Scholar]
- 55.Perez E, et al. Discovering language model behaviors with model-written evaluations. arXiv [csCL] 2022 [Google Scholar]
- 56.Nickerson RS. Confirmation bias: A ubiquitous phenomenon in many guises. Rev Gen Psychol. 1998;2:175–220. [Google Scholar]
- 57.Hart W, et al. Feeling validated versus being correct: A meta-analysis of selective exposure to information. Psychological Bulletin. 2009;135:555–588. doi: 10.1037/a0015701. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 58.McPherson M, Smith-Lovin L, Cook JM. Birds of a feather: Homophily in social networks. Annu Rev Sociol. 2001;27:415–444. [Google Scholar]
- 59.Kunda Z. The case for motivated reasoning. Psychological Bulletin. 1990;108:480–498. doi: 10.1037/0033-2909.108.3.480. [DOI] [PubMed] [Google Scholar]
- 60.Molden DC, Higgins ET. Motivated Thinking. Oxford University Press; 2012. [Google Scholar]
- 61.Vaswani A, et al. Attention is all you need. arXiv [csCL] 2017 [Google Scholar]
- 62.Devlin J, Chang M-W, Lee K, Toutanova K. BERT: Pre-training of deep bidirectional Transformers for language understanding. arXiv [csCL] 2018 [Google Scholar]
- 63.Radford A, Narasimhan K. Improving language understanding by generative pre-training. 2018 [Google Scholar]
- 64.Wolfram Stephen. What Is ChatGPT Doing … and Why Does It Work? 2023 https://writings.stephenwolfram.com/2023/02/what-is-chatgpt-doing-and-why-does-it-work/ [Google Scholar]
- 65.Gao Y, et al. Retrieval-Augmented Generation for Large Language Models: A Survey. 2024 doi: 10.48550/arXiv.2312.10997. [DOI] [Google Scholar]
- 66.Fountas Z, et al. Human-like episodic memory for infinite context LLMs. arXiv [csAI] 2024 [Google Scholar]
- 67.Christiano P, et al. Deep reinforcement learning from human preferences. arXiv [statML] 2017 [Google Scholar]
- 68.Ouyang L, et al. Training language models to follow instructions with human feedback. arXiv [csCL] 2022 [Google Scholar]
- 69.Zhang S, et al. Instruction tuning for large language models: A survey. arXiv [csCL] 2023 [Google Scholar]
- 70.Bai Y, et al. Constitutional AI: Harmlessness from AI Feedback. arXiv [csCL] 2022 [Google Scholar]
- 71.Gabriel S, Puri I, Xu X, Malgaroli M, Ghassemi M. Can AI relate: Testing large language model response for mental health support. arXiv [csCL] 2024 [Google Scholar]
- 72.OpenAI. Sycophancy in GPT-4o: what happened and what we’re doing about it. 2025. https://openai.com/index/sycophancy-in-gpt-4o/
- 73.Hartmann D, Wang SM, Pohlmann L, Berendt B. A systematic review of echo chamber research: comparative analysis of conceptualizations, operationalizations, and varying outcomes. J Comput Soc Sc. 2025;8:1–59. [Google Scholar]
- 74.Rathje S, et al. Sycophantic AI increases attitude extremity and overconfidence. OSF. 2025 [Google Scholar]
- 75.Anthropic. Agentic Misalignment: How LLMs could be insider threats. 2025. https://www.anthropic.com/research/agentic-misalignment .
- 76.Cloud A, et al. Subliminal Learning: Language models transmit behavioral traits via hidden signals in data. arXiv [csLG] 2025 [Google Scholar]
- 77.Chen R, Arditi A, Sleight H, Evans O, Lindsey J. Persona vectors: Monitoring and controlling character traits in language models. arXiv [csCL] 2025 [Google Scholar]
- 78.Ibrahim L, Sofia HF, Rocher L. Training language models to be warm and empathetic makes them less reliable and more sycophantic. arXiv [csCL] 2025 [Google Scholar]
- 79.Kumar A, Clune J, Lehman J, Stanley KO. Questioning representational optimism in deep learning: The fractured entangled representation hypothesis. arXiv [csCV] 2025 [Google Scholar]
- 80.Olah C, et al. Zoom In: An Introduction to Circuits. Distill. 2020;5:e00024–001. [Google Scholar]
- 81.Elhage N, et al. A Mathematical Framework for Transformer Circuits. 2021 https://transformer-circuits.pub/2021/framework/index.html . [Google Scholar]
- 82.Lanham T, et al. Measuring faithfulness in Chain-of-Thought reasoning. arXiv [csAI] 2023 [Google Scholar]
- 83.Turpin M, Michael J, Perez E, Bowman SR. Language Models don’t always say what they think: Unfaithful explanations in chain-of-thought prompting. arXiv [csCL] 2023 [Google Scholar]
- 84.Shojaee P, et al. The illusion of thinking: Understanding the strengths and limitations of reasoning models via the lens of problem complexity. arXiv [csAI] 2025 [Google Scholar]
- 85.Bo JY, Kazemitabaar M, Deng M, Inzlicht M, Anderson A. Invisible saboteurs: Sycophantic LLMs mislead novices in problem-solving tasks. arXiv [csHC] 2025 [Google Scholar]
- 86.Ibrahim L, et al. Multi-turn evaluation of anthropomorphic behaviours in large language models. arXiv [csCL] 2025 [Google Scholar]
- 87.Shaver PR, Mikulincer M. Attachment-related psychodynamics. Attach Hum Dev. 2002;4:133–161. doi: 10.1080/14616730210154171. [DOI] [PubMed] [Google Scholar]
- 88.Colombatto C, Birch J, Fleming SM. The influence of mental state attributions on trust in large language models. Commun Psychol. 2025;3:84. doi: 10.1038/s44271-025-00262-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 89.Hackenburg K, et al. The levers of political persuasion with conversational AI. arXiv [csCL] 2025 doi: 10.1126/science.aea3884. [DOI] [PubMed] [Google Scholar]
- 90.Glickman M, Sharot T. How human-AI feedback loops alter human perceptual, emotional and social judgements. Nat Hum Behav. 2025;9:345–359. doi: 10.1038/s41562-024-02077-2. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 91.Wang R, et al. PATIENT-\psi: Using large language models to simulate patients for training mental health professionals. arXiv [csCL] 2024 [Google Scholar]
- 92.Park JS, et al. Generative agent simulations of 1,000 people. arXiv [csAI] 2024 [Google Scholar]
- 93.Zheng L, et al. Judging LLM-as-a-judge with MT-bench and Chatbot Arena. arXiv [csCL] 2023 [Google Scholar]
- 94.Gibbs-Dean T, et al. Belief updating in psychosis, depression and anxiety disorders: A systematic review across computational modelling approaches. Neurosci Biobehav Rev. 2023;147:105087. doi: 10.1016/j.neubiorev.2023.105087. [DOI] [PubMed] [Google Scholar]
- 95.Adams RA, Huys QJM, Roiser JP. Computational Psychiatry: towards a mathematically informed understanding of mental illness. J Neurol Neurosurg Psychiatry. 2016;87:53–63. doi: 10.1136/jnnp-2015-310737. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 96.Huys QJM, Maia TV, Frank MJ. Computational psychiatry as a bridge from neuroscience to clinical applications. Nat Neurosci. 2016;19:404–413. doi: 10.1038/nn.4238. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 97.Dudley R, Taylor P, Wickham S, Hutton P. Psychosis, delusions and the ‘jumping to conclusions’ reasoning bias: A systematic review and meta-analysis. Schizophr Bull. 2016;42:652–665. doi: 10.1093/schbul/sbv150. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 98.McLean BF, Mattiske JK, Balzan RP. Association of the jumping to conclusions and evidence integration biases with delusions in psychosis: A detailed meta-analysis. Schizophr Bull. 2017;43:344–354. doi: 10.1093/schbul/sbw056. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 99.Sterzer P, et al. The predictive coding account of psychosis. Biol Psychiatry. 2018;84:634–643. doi: 10.1016/j.biopsych.2018.05.015. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 100.Clutterbuck RA, et al. Anthropomorphic tendencies in autism: A conceptual replication and extension of White and Remington (2019) and preliminary development of a novel anthropomorphism measure. Autism. 2022;26:940–950. doi: 10.1177/13623613211039387. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 101.Papadopoulos C. The use of AI chatbots for autistic people: A double-edged sword of digital support and companionship. Neurodiversity. 2025;3 [Google Scholar]
- 102.Kirk HR, Vidgen B, Röttger P, Hale SA. The benefits, risks and bounds of personalizing the alignment of large language models to individuals. Nat Mach Intell. 2024;6:383–392. [Google Scholar]
- 103.Wang J, et al. Social isolation in mental health: a conceptual and methodological review. Soc Psychiatry Psychiatr Epidemiol. 2017;52:1451–1461. doi: 10.1007/s00127-017-1446-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 104.Wickramaratne PJ, et al. Social connectedness as a determinant of mental health: A scoping review. PLoS One. 2022;17:e0275004. doi: 10.1371/journal.pone.0275004. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 105.Epley N, Akalis S, Waytz A, Cacioppo JT. Creating social connection through inferential reproduction: loneliness and perceived agency in gadgets, gods, and greyhounds: Loneliness and perceived agency in gadgets, gods, and greyhounds. Psychol Sci. 2008;19:114–120. doi: 10.1111/j.1467-9280.2008.02056.x. [DOI] [PubMed] [Google Scholar]
- 106.Varadarajan V, et al. The consistent lack of variance of psychological factors expressed by LLMs and spambots. 2025:111–119. [Google Scholar]
- 107.Shumailov I, et al. AI models collapse when trained on recursively generated data. Nature. 2024;631:755–759. doi: 10.1038/s41586-024-07566-y. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 108.Yeung JA, Dalmasso J, Foschini L, Dobson RJB, Kraljevic Z. The psychogenic machine: Simulating AI psychosis, delusion reinforcement and Harm Enablement in large language models. arXiv [csLG] 2025 [Google Scholar]
- 109.Stein F, et al. Transdiagnostic types of formal thought disorder and their association with gray matter brain structure: a model-based cluster analytic approach. Mol Psychiatry. 2025;30:4286–4295. doi: 10.1038/s41380-025-03009-w. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 110.Kircher T, Bröhl H, Meier F, Engelen J. Formal thought disorders: from phenomenology to neurobiology. Lancet Psychiatry. 2018;5:515–526. doi: 10.1016/S2215-0366(18)30059-2. [DOI] [PubMed] [Google Scholar]
- 111.Garcia B, Chua EYS, Brah HS. The problem of atypicality in LLM-powered psychiatry. J Med Ethics. 2025 doi: 10.1136/jme-2025-110972. [DOI] [PubMed] [Google Scholar]
- 112.Anil C, et al. Many-shot Jailbreaking. Advances in Neural Information Processing Systems. 2024;37 [Google Scholar]
- 113.Summerfield C, et al. The impact of advanced AI systems on democracy. Nat Hum Behav. 2025 doi: 10.1038/s41562-025-02309-z. [DOI] [PubMed] [Google Scholar]
- 114.Shen H, et al. Towards Bidirectional Human-AI alignment: A systematic review for clarifications, framework, and future directions. arXiv [csHC] 2024 [Google Scholar]


