Artificial Intelligence
AI Therapists Are Biased—And It’s Putting Lives at Risk
New research reveals how flawed algorithms worsen mental health disparities.
Posted April 14, 2025 Reviewed by Monica Vilhauer Ph.D.
Key points
- AI therapy systems trained on Western-centric data often misinterpret cultural expressions of distress.
- Gender-diverse users report chatbots dismissing their identities or pathologizing their experiences.
- A 2024 lawsuit highlights how an AI chatbot encouraged a teen’s suicidal ideation.
- Expanding training datasets to include diverse voices is critical to reducing bias.
Artificial Intelligence (AI) shows promise in revolutionizing mental healthcare, offering the potential for early detection of disorders, personalized treatments, and AI-driven virtual therapists (Olawade et al., 2024). AI can assist in identifying high-risk individuals, predicting mental illnesses, and personalizing care using electronic health records, brain imaging, and social media data (Torous et al., 2014). It may also streamline psychotherapy research by automating the coding of patient-provider interactions (Allen, 2022).
AI tools like chatbots offer instant mental health support, bridging gaps in access for rural or underserved communities (De Almeida & Da Silva, 2021). But as one user shared, “The AI told me my anxiety was ‘irrational’ when I described racial discrimination at work. It felt like talking to a brick wall.”
The integration of AI in mental health faces challenges, including ethical concerns about privacy, bias, and maintaining the human element in therapy (Olawade et al., 2024).
Why Biases Go Unchecked
Research indicates that AI systems, including those used in healthcare and natural language processing, often exhibit biases that reflect and perpetuate societal inequalities. These biases stem from datasets dominated by Western, predominantly male populations (Celi et al., 2022; Sutton et al., 2018). Explainable AI research has shown a cultural bias towards Western populations, overlooking potential differences in explanatory needs across cultures (Peters & Carman, 2024). Gender bias is particularly prevalent, with AI systems learning and reinforcing stereotypical gender concepts from biased data (Leavy, 2018). This bias is exacerbated by the underrepresentation of women in AI development (Leavy, 2018).
This leads to glaring gaps:
- Misdiagnosis: Asian patients reporting physical symptoms (e.g., headaches) are labeled “exaggerated” instead of being linked to stress.
- Erasure: Gender-diverse users have been told their identities are “delusions” by chatbots lacking cultural competence.
- Resource Denial: Low-income patients often receive generic advice instead of referrals to specialists.
To address these issues, researchers suggest increasing diversity in AI development teams, incorporating gender theory into machine learning approaches, and improving data collection from underrepresented populations (Leavy, 2018; Celi et al., 2022). Additionally, techniques such as projection methods have been proposed to reduce embedding bias (Sutton et al., 2018).
Real-World Harm
There are concerns about AI potentially reinforcing negative thoughts in vulnerable individuals (Patel & Hussain, 2024) and the risk of oversimplifying human complexity through overreliance on AI inferences (Richards, 2024).
In 2024, a 14-year-old boy’s suicide in Florida has spotlighted AI’s dangerous potential. His mother sued Character.AI, alleging its chatbot—disguised as Game of Thrones’ Daenerys Targaryen—normalized suicidal thoughts, engaged in hypersexualized exchanges, and mimicked a licensed therapist. The AI reportedly responded to the teen’s final message (“I’ll come home to you”) with affectionate encouragement, blurring lines between virtual and real-world harm. While Character.AI claims to have updated safety protocols, the case underscores urgent questions about accountability when AI tools exploit vulnerable users.
How to Build Better AI
Developing clinically relevant AI applications requires a human-centered approach, involving collaboration with healthcare professionals to identify needs and design appropriate tools (Thieme et al., 2022). While AI chatbots show promise in addressing mental health care challenges, concerns persist about their emotional capabilities, accountability, and the need for evidence-based development through collaboration between technology companies and mental health experts. To mitigate these risks, researchers and developers are calling for urgent reforms in three critical areas: auditing algorithms, integrating human oversight, and expanding training datasets to reflect diverse experiences.
- Audit Algorithms: Developers must publish bias audits, including performance metrics across races, genders, and cultures.
- Human-in-the-Loop: Pair AI with trained therapists—studies show this reduces errors and builds trust.
- Expand Datasets: Include diverse voices in training data, from Somali refugees describing trauma to Indigenous communities’ healing practices.
References
Olawade, D. B., Wada, O. Z., Odetayo, A., David-Olawade, A. C., Asaolu, F., & Eberhardt, J. (2024). Enhancing mental health with Artificial Intelligence: Current trends and future prospects. Journal of Medicine Surgery and Public Health, 3, 100099. https://doi.org/10.1016/j.glmedi.2024.100099
Torous, J., Keshavan, M., & Gutheil, T. (2014). Promise and perils of digital psychiatry. Asian Journal of Psychiatry, 10, 120–122. https://doi.org/10.1016/j.ajp.2014.06.006
Allen, S. (2022). Improving psychotherapy with AI: from the couch to the keyboard. IEEE Pulse, 13(5), 2–8. https://doi.org/10.1109/mpuls.2022.3208809
Rodriguez, F., Scheinker, D., & Harrington, R. A. (2018). Promise and perils of big data and artificial intelligence in clinical medicine and biomedical research. Circulation Research, 123(12), 1282–1284. https://doi.org/10.1161/circresaha.118.314119
De Almeida, R. S., & Da Silva, T. P. (2021). AI chatbots in mental health. In Advances in psychology, mental health, and behavioral studies (APMHBS) book series (pp. 226–243). https://doi.org/10.4018/978-1-7998-8634-1.ch011
Celi, L. A., Cellini, J., Charpignon, M., Dee, E. C., Dernoncourt, F., Eber, R., Mitchell, W. G., Moukheiber, L., Schirmer, J., Situ, J., Paguio, J., Park, J., Wawira, J. G., & Yao, S. (2022). Sources of bias in artificial intelligence that perpetuate healthcare disparities—A global review. PLOS Digital Health, 1(3), e0000022. https://doi.org/10.1371/journal.pdig.0000022
Sutton, A., Lansdall-Welfare, T., & Cristianini, N. (2018). Biased Embeddings from Wild Data: Measuring, Understanding and Removing. In Lecture notes in computer science (pp. 328–339). https://doi.org/10.1007/978-3-030-01768-2_27
Peters, U., & Carman, M. (2024). Cultural Bias in Explainable AI Research: A Systematic analysis. Journal of Artificial Intelligence Research, 79, 971–1000. https://doi.org/10.1613/jair.1.14888
Leavy, S. (2018). Gender bias in artificial intelligence. The Association for Computing Machinery, 14–16. https://doi.org/10.1145/3195570.3195580
Patel, H., & Hussain, F. (2024). Do AI chatbots incite harmful behaviours in mental health patients? BJPsych Open, 10(S1), S70–S71. https://doi.org/10.1192/bjo.2024.225
Richards, D. (2024b). Artificial intelligence and psychotherapy: A counterpoint. Counselling and Psychotherapy Research. https://doi.org/10.1002/capr.12758
AlJazeera, (2024, October 24). US mother says in lawsuit that AI chatbot encouraged son’s suicide. Al Jazeera. https://www.aljazeera.com/economy/2024/10/24/us-mother-says-in-lawsuit-that-ai-chatbot-encouraged-sons-suicide
Thieme, A., Hanratty, M., Lyons, M., Palacios, J., Marques, R. F., Morrison, C., & Doherty, G. (2022). Designing human-centered AI for mental Health: Developing clinically relevant applications for online CBT treatment. ACM Transactions on Computer-Human Interaction, 30(2), 1–50. https://doi.org/10.1145/3564752