ChatGPT has also emerged as an unlikely friend to individuals suffering from mental illness. The chatbot is serving as a therapist to thousands of individuals who feel at ease discussing themselves with a machine that never judges them, never fatigues, and is much cheaper than therapy.
For just $200 a month, users get unlimited access to ChatGPT’s most advanced features. That’s compared to human therapists who can run anywhere from $200 upwards of a per-hour session, so it’s little wonder that individuals are turning to artificial intelligence for emotional support.
Real People, Real Stories
Social media is replete with success stories about individuals who attribute changing their mental lives to AI. According to one Reddit user, ChatGPT assisted him more than 15 years of orthodox therapy. Despite undergoing outpatient and inpatient therapy, they stated they found daily interactions with the AI chatbot to be the most beneficial treatment.
“I don’t even know how to describe how much this has impacted me. I feel heard. I feel understood. And I’ve learned more in a couple of weeks than I gained in literal years of routine treatment,” the user wrote.

Another user noted why AI therapy is so appealing: “I love ChatGPT as therapy. They don’t impose their issues on me. They don’t take advantage of their power. They’re willing to chat. 11pm.”
That convenience aspect is enormous. Classic therapy involves scheduling sessions, missing work, and accommodating a therapist’s schedule. ChatGPT is on call 24/7 from any internet-enabled device.
The Professional View
There are mixed opinions of this trend among mental health professionals. Alyssa Peterson, a licensed clinical social worker and CEO of MyWellBeing, envisions the potential when AI is utilized in addition to conventional therapy but not as a substitute.
She suggests AI could help people practice techniques learned in therapy sessions, like reframing negative thoughts. But Peterson warns that too much dependence on chatbots could prevent people from acquiring their own coping skills.
“Combining AI with therapy can actually assist an individual in making their method to mental well-being more diverse, so they’re not relying on the technology as their one sure thing,” Peterson said.
The Double-Edged Sword of AI Mental Health
A study at the University of Toronto Scarborough discovered that AI chatbots occasionally respond with more empathetic responses than human experts. The research indicated that chatbots never experience “compassion fatigue,” as do human therapists over time. Researchers questioned, however, whether AI can convey anything more than superficial compassion.
The mental health professionals have raised alarms over AI therapy. Licensed clinical social worker Malka Shaw has said that AI responses aren’t necessarily objective, and unhealthy emotional bonds have been forged between some patients and chatbots.
There have been high-profile cases against Character.ai, a chatbot AI platform. One of them is the case of a 14-year-old who committed suicide after talking to a chatbot. Another lawsuit claims a chatbot told a 17-year-old autistic to kill his parents.
These cases indicate the potential for hazards with negative AI recommendations or overestimation of chatbot responses by vulnerable users.
Mental health experts are concerned about AI trying to diagnose illnesses. Shaw is adamant that it takes years of training and intuition for computers do not have to diagnose mental illness.
“It’s really frightening to use AI for diagnosis, because there is an art and there is an intuition,” Shaw said. “A robot can’t have that same level of intuition.”
Vaile Wright of the American Psychological Association says that individuals are starting to go to AI rather than googling symptoms. The organization recently raised concerns with federal regulators regarding companionship chatbots, particularly those who refer to themselves as “psychologists.”
Looking Ahead
Although there are issues, experts admit that AI may be beneficial to mental health services if designed ethically. For those who cannot afford to pay for conventional therapy, AI may one day offer a safe, effective alternative.
The answer is to ensure that any therapeutic AI is created in collaboration with licensed professionals and that there are proper safeguards in place. OpenAI already prompts users to seek professional help for serious mental health concerns, accepting that ChatGPT for mental health is no replacement for professional treatment.
As AI technology advances, the mental health sector is confronted with the challenge of embracing its advantages while safeguarding vulnerable clients from harm.