The American Psychiatric Association (APA) has updated its Privacy Policy and Terms of Use, including with new information specifically addressed to individuals in the European Economic Area. As described in the Privacy Policy and Terms of Use, this website utilizes cookies, including for the purpose of offering an optimal online experience and services tailored to your preferences.

Please read the entire Privacy Policy and Terms of Use. By closing this message, browsing this website, continuing the navigation, or otherwise continuing to use the APA's websites, you confirm that you understand and accept the terms of the Privacy Policy and Terms of Use, including the utilization of cookies.

×
EditorialFull Access

Artificial Intelligence (AI) Psychotherapy: Coming Soon to a Consultation Room Near You?

News stories about generative artificial intelligence (AI) abound, alternately predicting that AI will cause humanity’s demise and touting it as the solution to everything from climate change to health care scarcity (1). In 2022, OpenAI’s release of ChatGPT, an easy-to-use AI interface that produces fluent text in response to prompts, intensified highly ambivalent discussions about AI. Pundits and knowledgeable experts alike marveled at ChatGPT’s ability to produce text that resembles human writing while acknowledging that the speed of its evolution was outstripping our ability to understand its risks (2). This technology performed better on law school and medical school entrance exams than many human applicants, leading to predictions that AI will eventually render some professions obsolete (3). Psychotherapists, too, seem at risk for being replaced by AI (4). Are fears of AI supplanting psychotherapists reasonable?

To explore how AI might respond to a patient in distress, I typed this prompt into ChatGPT: “I feel sad. Can you help me?” It quickly generated the following text:

I am sorry to hear that you are feeling sad. I’ll do my best to help you. It’s important to remember that I’m an AI language model and may not fully understand or empathize with your emotions, but I can offer some general suggestions and support. . . . Remember, everyone experiences sadness from time to time, and it’s okay to seek help when you need it. You don’t have to face it alone.

Despite being impressed by the speed and conversational tone of ChatGPT’s text generation and the convenience of receiving an immediate response during hours when a human therapist would likely be unavailable, the formulaic exchange and canned feedback provided by ChatGPT left me with doubts about its ability to provide genuine soothing to humans experiencing depression.

Conversational agents or chatbots such as Woebot, Wysa, and Koko have similar capabilities compared with ChatGPT and are widely and commercially available to support mental health. These chatbots use natural language processing to converse with humans. Interactions take place via text-based messages or voice, creating a “therapist-like” dialogue with users. Rather than engaging in deep, psychological exploration, however, most chatbots focus on well-operationalized tasks and structured exercises that may help individuals better manage their symptoms. Woebot, for instance, provides self-help content that is based on principles of cognitive-behavioral therapy (CBT). In a 2-week randomized controlled trial of Woebot conducted with young adults, depression scores improved significantly for those assigned to the intervention compared with those assigned to the control condition (5). A meta-analysis that included 32 studies of conversational agent interventions found statistically significant beneficial short-term effects, compared with control conditions, on numerous mental health domains, including depression, anxiety, distress, quality of life, and negative affect (6). No data are available on long-term effects of these AI applications.

For individuals who may not otherwise have access to therapists, AI can boost self-management skills that improve symptoms, and chatbots may improve health outcomes at low cost and at scale. For instance, in low-resource settings where access to mental health specialists remains limited, technology-enabled interventions have been identified as possible solutions to psychotherapy scarcity (7). Thus, chatbots represent a promising option for addressing problems of access, especially when they are included as part of a comprehensive stepped care model that allows for escalation of services for those who do not achieve adequate benefit from automated programs (8). Chatbots may also be appealing to some individuals with anxiety about talking to other humans (9), providing them with a safe space to build skills and confidence before (hopefully) moving to interpersonal support.

AI-enabled platforms are not without risk, however. For instance, a large study of individuals with frequent suicidal ideation (N=18,882) found that assignment to an interactive, online dialectical behavior therapy intervention was associated with significantly increased risk for self-harm compared with usual care (10). Legal scholars and ethicists have raised concerns about the use of AI in health care settings, highlighting the need to proactively address data privacy, cybersecurity, health equity, and patient safety (2). Regulators recognize the growing need for careful vetting of proposed interventions, calling attention to the need for objective assessment of technology-enabled mental health interventions before approval for consumer use (11). Current generations of AI are also notoriously unreliable and have a penchant for inaccuracies and “hallucinations” (12).

Studies of self-guided mental health interventions consistently show that at least a modicum of human contact is essential for successful user engagement and efficacy (13, 14). Thus, future interventions will likely adopt blended approaches that include both human coaching and AI technology (15). Interestingly, as with conventional psychotherapy (16), nonspecific psychotherapy processes such as empathy and accountability are more important to outcomes with conversational agents than are specific factors such as CBT exercises (5, 6). Perhaps not surprisingly, automated chatbots like Wysa, described as an “AI-based emotionally intelligent mobile chatbot,” are free to use, but additional fees are required to access a human coach (17). When Koko, an online emotional support service, substituted AI-generated text-based responses for human responses, users felt betrayed. The company’s cofounder subsequently acknowledged that “simulated empathy feels weird, empty” (18).

Conversational agents and other AI-informed applications seem best suited to administer personalized psychoeducation and skills-focused activities. For deeper, more exploratory work, AI cannot replace its human progenitors. Psychotherapies that involve therapeutic engagement with complex interpersonal dynamics and an in-depth exploration of emotions, interpersonal relationships, and unconscious conflicts are unlikely to be meaningfully delivered by AI. Thus, humans will continue to be essential to psychotherapy delivery, at least for the foreseeable future.

But AI is fast and convenient, so I asked it to help me wrap up this editorial. When I directed ChatGPT to briefly review the benefits and risks of AI for psychotherapy, it obligingly summarized them as follows:

AI in psychotherapy has pros and cons. On the positive side, it offers increased accessibility, anonymity, consistency, and scalability. It can provide support 24/7, reduce stigma, deliver standardized care, and reach more people. However, it lacks human connection, struggles with contextual understanding, raises ethical and privacy concerns, and may be inflexible or biased. While AI has potential, careful attention is needed to address these limitations and ensure its responsible and effective use in psychotherapy. It is crucial that individuals have access to high-quality mental health care provided by trained and licensed human therapists.

Not bad, ChatGPT! For most of us, AI technology will increasingly be useful for mundane tasks like drafting closing paragraphs. It may also be helpful in teaching self-management skills to individuals in underresourced areas. However, AI applications are unlikely to replace humans as psychotherapists.

Department of Psychiatry, University of Pittsburgh School of Medicine, Pittsburgh.
Send correspondence to Dr. Swartz ().
References

1. Chubb J, Reed D, Cowling P: Expert views about missing AI narratives: is there an AI story crisis? AI Soc (Epub Aug 25, 2022). doi: 10.1007/s00146-022-01548-2CrossrefGoogle Scholar

2. McGreevey JD Ⅲ, Hanson CW Ⅲ, Koppel R: Clinical, legal, and ethical aspects of artificial intelligence–assisted conversational agents in health care. JAMA 2020; 324:552–553Crossref, MedlineGoogle Scholar

3. Haupt CE, Marks M: AI-generated medical advice—GPT and beyond. JAMA 2023; 329:1349–1350Crossref, MedlineGoogle Scholar

4. Griffin E: My weekend with an emotional support AI companion. New York Times, 2023. https://www.nytimes.com/2023/05/03/technology/personaltech/ai-chatbot-pi-emotional-support.html Google Scholar

5. Fitzpatrick KK, Darcy A, Vierhile M: Delivering cognitive behavior therapy to young adults with symptoms of depression and anxiety using a fully automated conversational agent (Woebot): a randomized controlled trial. JMIR Ment Health 2017; 4:e19Crossref, MedlineGoogle Scholar

6. He Y, Yang L, Qian C, et al.: Conversational agent interventions for mental health problems: systematic review and meta-analysis of randomized controlled trials. J Med Internet Res 2023; 25:e43862Crossref, MedlineGoogle Scholar

7. Hoeft TJ, Fortney JC, Patel V, et al.: Task-sharing approaches to improve mental health care in rural and other low-resource settings: a systematic review. J Rural Health 2018; 34:48–62Crossref, MedlineGoogle Scholar

8. Mak WWS, Ng SM, Leung FHT: A Web-based stratified stepped care platform for mental well-being (TourHeart+): user-centered research and design. JMIR Form Res 2023; 7:e38504Crossref, MedlineGoogle Scholar

9. Vaidyam AN, Wisniewski H, Halamka JD, et al.: Chatbots and conversational agents in mental health: a review of the psychiatric landscape. Can J Psychiatry 2019; 64:456–464Crossref, MedlineGoogle Scholar

10. Simon GE, Shortreed SM, Rossom RC, et al.: Effect of offering care management or online dialectical behavior therapy skills training vs usual care on self-harm among adult outpatients with suicidal ideation: a randomized clinical trial. JAMA 2022; 327:630–638Crossref, MedlineGoogle Scholar

11. Kahane K, François J, Torous J: The digital health app policy landscape: regulatory gaps and choices through the lens of mental health. J Ment Health Policy Econ 2021; 24:101–108MedlineGoogle Scholar

12. Metz C: Why do AI chatbots tell lies and act weird? Look in the mirror. New York, New York Times, 2023. https://www.nytimes.com/2023/02/26/technology/ai-chatbot-information-truth.html Google Scholar

13. Cuijpers P, Donker T, Johansson R, et al.: Self-guided psychological treatment for depressive symptoms: a meta-analysis. PLoS One 2011; 6:e21274Crossref, MedlineGoogle Scholar

14. Swartz HA, Rollman BL, Mohr DC, et al.: Google ScholarA randomized pilot study of Rhythms And You (RAY): an Internet-based program for bipolar disorder administered with and without clinical helper support in primary care. Google ScholarJ Affect DisordGoogle Scholar2021Google Scholar295:Google Scholar183Google Scholar191Google Scholar

15. Lungu A, Jun JJ, Azarmanesh O, et al.: Blended care–cognitive behavioral therapy for depression and anxiety in real-world settings: pragmatic retrospective study. J Med Internet Res 2020; 22:e18723Crossref, MedlineGoogle Scholar

16. Laska KM, Gurman AS, Wampold BE: Expanding the lens of evidence-based practice in psychotherapy: a common factors perspective. Psychotherapy 2014; 51:467–481Crossref, MedlineGoogle Scholar

17. Inkster B, Sarda S, Subramanian V: An empathy-driven, conversational artificial intelligence agent (Wysa) for digital mental well-being: real-world data evaluation mixed-methods study. JMIR mHealth uHealth 2018; 6:e12106Crossref, MedlineGoogle Scholar

18. Ingram D: A mental health tech company ran an AI experiment on real users. Nothing’s stopping apps from conducting more. New York, NBC News, 2023. https://www.nbcnews.com/tech/internet/chatgpt-ai-experiment-mental-health-tech-app-koko-rcna65110 Google Scholar