In the pre-dawn hours, Ann Li’s anxieties felt overwhelming. She’d recently been diagnosed with a serious health problem, and she just wanted to talk to someone about it. But she hadn’t told her family, and all her friends were asleep. So instead, she turned to ChatGPT.
“It’s easier to talk to AI during those nights,” said the 30-year-old Taiwanese woman.
In China, Yang, a 25-year-old Guangdong resident, had never seen a mental health professional when she started talking to an AI chatbot earlier this year. Yang says it was difficult to access mental health services, and she couldn’t contemplate confiding in family or friends.
Photo: Reuters
“Telling the truth to real people feels impossible,” she says.
But she was soon talking to the chatbot “day and night.”
Li and Yang are among a growing number of Chinese-speaking people turning to generative AI chatbots instead of professional human therapists. Experts say there is huge potential for AI in the mental health sector, but are concerned about the risks of people in distress turning to the technology, rather than human beings, for medical assistance.
There are few official statistics, but mental health professionals in Taiwan and China have reported rising rates of patients consulting AI before seeing them, or instead of seeing them. Surveys, including a global analysis recently published by Harvard Business Review, show psychological assistance is now a leading reason for adults to use AI chatbots. On social media there are hundreds of thousands of posts praising AI for helping them.
It comes amid rising rates of mental illness in Taiwan and China, particularly among younger people. Access to services is not keeping apace — appointments are hard to get, and they’re expensive. Chatbot users say AI saves them time and money, gives real answers and is more discrete in a society where there is still stigma around mental health.
“In some way the chatbot does help us — it’s accessible, especially when ethnic Chinese tend to suppress or downplay our feelings,” says Dr Yi-Hsien Su, a clinical psychologist at True Colors in Taiwan, who also works in schools and hospitals to promote mental wellbeing in Taiwan.
“I talk to people from Gen Z and they’re more willing to talk about problems and difficulties … But there’s still much to do.”
In Taiwan, the most popular chatbot is ChatGPT. In China, where western apps like ChatGPT are banned, people have turned to domestic offerings like Baidu’s Ernie Bot, or the recently launched DeepSeek. They are all advancing at rapid speed, and are incorporating wellbeing and therapy into responses as demand increases.
User experiences vary. Li says ChatGPT gives her what she wants to hear, but that can also be predictable and uninsightful. She also misses the process of self discovery in counseling.
“I think AI tends to give you the answer, the conclusion that you would get after you finish maybe two or three sessions of therapy,” she says.
Yet 27-year-old Nabi Liu, a Taiwanese woman based in London, has found the experience to be very fulfilling.
“When you share something with a friend, they might not always relate. But ChatGPT responds seriously and immediately,” she says. “I feel like it’s genuinely responding to me each time.”
Experts say it can assist people who are in distress but perhaps don’t need professional help yet, like Li, or those who need a little encouragement to take the next step.
Yang says she doubted whether her struggles were serious enough to warrant professional help.
“Only recently have I begun to realize that I might actually need a proper diagnosis at a hospital,” she says.
“Going from being able to talk [to AI] to being able to talk to real people might sound simple and basic, but for the person I was before, it was unimaginable.”
But experts have also raised concerns about people falling through the cracks, missing the signs that Yang saw for herself, and not getting the help they need.
There have been tragic cases in recent years of young people in distress seeking help from chatbots instead of professionals, and later taking their own lives.
“AI mostly deals with text, but there are things we call non verbal input. When a patient comes in maybe they act differently to how they speak but we can recognize those inputs,” Su says.
A spokesperson for the Taiwan Counseling Psychology Association says AI can be an “auxiliary tool,” but couldn’t replace professional assistance “let alone the intervention and treatment of psychologists in crisis situations.”
“AI has the potential to become an important resource for promoting the popularization of mental health. However, the complexity and interpersonal depth of the clinical scene still require the real ‘present’ psychological professional.”
The association says AI can be “overly positive,” miss cues and delay necessary medical care. It also operates outside the peer review and ethics codes of the profession.
“In the long run, unless AI develops breakthrough technologies beyond current imagination, the core structure of psychotherapy should not be shaken.”
Su says he’s excited about the ways AI could modernize and improve his industry, noting potential uses in training of professionals and detecting people online who might need intervention. But for now he recommends people approach the tools with caution.
“It’s a simulation, it’s a good tool, but has limits and you don’t know how the answer was made,” he says.
Behind a car repair business on a nondescript Thai street are the cherished pets of a rising TikTok animal influencer: two lions and a 200-kilogram lion-tiger hybrid called “Big George.” Lion ownership is legal in Thailand, and Tharnuwarht Plengkemratch is an enthusiastic advocate, posting updates on his feline companions to nearly three million followers. “They’re playful and affectionate, just like dogs or cats,” he said from inside their cage complex at his home in the northern city of Chiang Mai. Thailand’s captive lion population has exploded in recent years, with nearly 500 registered in zoos, breeding farms, petting cafes and homes. Experts warn the
The unexpected collapse of the recall campaigns is being viewed through many lenses, most of them skewed and self-absorbed. The international media unsurprisingly focuses on what they perceive as the message that Taiwanese voters were sending in the failure of the mass recall, especially to China, the US and to friendly Western nations. This made some sense prior to early last month. One of the main arguments used by recall campaigners for recalling Chinese Nationalist Party (KMT) lawmakers was that they were too pro-China, and by extension not to be trusted with defending the nation. Also by extension, that argument could be
Aug. 4 to Aug. 10 When Coca-Cola finally pushed its way into Taiwan’s market in 1968, it allegedly vowed to wipe out its major domestic rival Hey Song within five years. But Hey Song, which began as a manual operation in a family cow shed in 1925, had proven its resilience, surviving numerous setbacks — including the loss of autonomy and nearly all its assets due to the Japanese colonial government’s wartime economic policy. By the 1960s, Hey Song had risen to the top of Taiwan’s beverage industry. This success was driven not only by president Chang Wen-chi’s
Last week, on the heels of the recall election that turned out so badly for Taiwan, came the news that US President Donald Trump had blocked the transit of President William Lai (賴清德) through the US on his way to Latin America. A few days later the international media reported that in June a scheduled visit by Minister of National Defense Wellington Koo (顧立雄) for high level meetings was canceled by the US after China’s President Xi Jinping (習近平) asked Trump to curb US engagement with Taiwan during a June phone call. The cancellation of Lai’s transit was a gaudy