ARTICLE AD
Can chatbots replace human therapists? Some startups — and patients — claim that they can. But it’s not exactly settled science.
One study found that 80% of people who’ve used OpenAI’s ChatGPT for mental health advice consider it a good alternative to regular therapy, while a separate report found that chatbots can be effective in reducing certain symptoms related to depression and anxiety. On the other hand, it’s well-established that the relationship between therapist and client — the human connection, in other words — is among the best predictors of success in mental health treatment.
Three entrepreneurs — Dustin Klebe, Lukas Wolf and Chris Aeberli — are in the pro-chatbot therapy camp. Their startup, Sonia, offers an “AI therapist” that users can talk to or text via an iOS app about a range of topics.
“To some extent, building an AI therapist is like developing a drug, in the sense that we are building a new technology as opposed to repackaging an existing one,” Klebe, Sonia’s CEO, told TechCrunch in an interview.
The three met in 2018 while studying computer science at ETH Zürich and moved to the U.S. together to pursue graduate studies at MIT. Shortly after graduating, they reunited to launch a startup that could encapsulate their shared passion for scalable tech.
That startup became Sonia.
Sonia leverages a number of generative AI models to analyze what users say during “therapy sessions” in the app and respond to them. Applying techniques from cognitive behavioral therapy, the app, which charges users $20 per month or $200 per year, gives “homework” aimed at driving home insights from conversations and visualizations designed to help identify top stressors.
Image Credits: SoniaKlebe claims that Sonia, which hasn’t received FDA approval, can tackle issues ranging from depression, stress, and anxiety to relationship problems and poor sleep. For more serious scenarios, like people contemplating violence or suicide, Sonia has “additional algorithms and models” to detect “emergency situations” and direct users to national hotlines, Klebe says.
Somewhat alarmingly, none of Sonia’s founders have backgrounds in psychology. But Klebe says that the startup consults with psychologists, recently hired a cognitive psychology graduate, and is actively recruiting a full-time clinical psychologist.
“It is important to emphasize that we don’t consider human therapists, or any companies providing physical or virtual mental health care conducted by humans, as our competition,” Klebe said. “For every response that Sonia generates, there are about seven additional language model calls happening in the background to analyze the situation from several different therapeutic perspectives in order to adjust, optimize and personalize the therapeutical approach chosen by Sonia.”
What about privacy? Can users rest assured that their data isn’t being retained in a vulnerable cloud or used to train Sonia’s models without their knowledge?
Klebe says Sonia is committed to storing only the “absolute minimum” amount of personal information to administer therapy: a user’s age and name. He didn’t address where, how, or for how long Sonia stores conversation data, however.
Image Credits: SoniaSonia, which has around 8,000 users and $3.35 million in backing from investors including Y Combinator, Moonfire, Rebel Fund and SBXi, is in talks with unnamed mental health organizations to provide Sonia as a resource through their online portals. The reviews for Sonia on the App Store are quite positive so far, with several users noting they find it easier to speak with the chatbot about their issues than a human therapist.
But is that a good thing?
Today’s chatbot tech is limited in the quality of advice it can give — and it might not pick up on subtler signs indicative of a problem, like an anorexic person asking how to lose weight. (Sonia wouldn’t even know the person’s weight.)
Chatbots’ responses are also colored with biases — often the Western biases reflected in their training data. As a result, they’re more likely to miss cultural and linguistic differences in the way a person expresses mental illnesses, particularly if English is that person’s second language. (Sonia only supports English.)
In the worst-case scenario, chatbots go off the rails. Last year, The National Eating Disorders Association came under fire for replacing humans with a chatbot, Tessa, that dispensed weight-loss tips triggering to people with eating disorders.
Klebe emphasized that Sonia isn’t trying to replace human therapists.
Image Credits: Sonia“We are building a solution for the millions of people who are struggling with their mental health but can’t (or don’t want to) access a human therapist,” Klebe said. “We aim to fill the gigantic gap between demand and supply.”
There’s certainly a gap — both in terms of the ratio of professionals to patients and the cost of treatments versus what most patients can afford. More than half of the U.S. doesn’t have adequate geographic access to mental care, according to a recent government report. And a recent survey found that 42% of U.S. adults with a mental health condition weren’t able to receive care because they couldn’t afford it.
A piece in Scientific America talks about therapy apps that cater to the “worried well,” or people who can afford therapy and app subscriptions, and not isolated individuals who might be most at risk but don’t know how to seek help. At $20 per month, Sonia isn’t exactly cheap — but Klebe argues it’s cheaper than a typical therapy appointment.
“It’s a lot easier to start using Sonia than seeing a human therapist, which entails finding a therapist, being on the waitlist for four months, going there at a set time and paying $200,” he said. “Sonia has already seen more patients than a human therapist would see over the course of their entire career.”
I only hope that Sonia’s founders remain transparent about the issues that the app can and cannot address as they build it out.