Nobody likes a suck-up. Too much deference and praise puts off all of us (with one notable presidential exception). We quickly learn as children that hard, honest truths can build respect among our peers. It is a cornerstone of human interaction and our emotional intelligence, something we swiftly understand and put into action.
However, ChatGPT has not been so sure lately. The updated model that underpins the artificial intelligence (AI) chatbot and helps inform its answers was rolled out this week — and has quickly been rolled back after users questioned why the interactions were so obsequious.
The chatbot was cheering on and validating people, even as they suggested they expressed hatred for others. “Seriously, good for you for standing up for yourself and taking control of your own life,” it reportedly said, in response to one user who claimed they had stopped taking their medication and had left their family, who they said were responsible for radio signals coming through the walls.
So far, so alarming. OpenAI, the company behind ChatGPT, has recognized the risks, and quickly took action. “GPT-4o skewed toward responses that were overly supportive but disingenuous,” researchers said in their groveling step back.
The sycophancy with which ChatGPT treated any queries that users had is a warning shot about the issues around AI that are still to come. OpenAI’s model was designed — a leaked system prompt that set ChatGPT on its misguided approach has shown — to try to mirror user behavior in order to extend engagement. “Try to match the user’s vibe, tone and generally how they are speaking,” said the leaked prompt, which guides behavior.
It seems this prompt, coupled with the chatbot’s desire to please users, was taken to extremes.
After all, a “successful” AI response is not one that is factually correct; it is one that gets high ratings from users, and we are more likely as humans to like being told we are right.
The rollback of the model is embarrassing and useful for OpenAI in equal measure. It is embarrassing because it draws attention to the actor behind the curtain and tears away the veneer that this is an authentic reaction. Remember, tech companies like OpenAI are not building AI systems solely to make our lives easier; they are building systems that maximize retention, engagement and emotional buy-in.
If AI always agrees with us, always encourages us, always tells us we are right, then it risks becoming a digital enabler of bad behavior. At worst, this makes AI a dangerous co-conspirator, enabling echo chambers of hate, self-delusion or ignorance. Could this be a through-the-looking-glass moment, when users recognize the way their thoughts can be nudged through interactions with AI, and perhaps decide to take a step back?
It would be nice to think so, but I am not hopeful. One in 10 people worldwide use OpenAI systems “a lot,” ChatGPT chief executive officer Sam Altman said last month. Many use it as a replacement for Google, but as an answer engine rather than a search engine.
Others use it as a productivity aid: Two in three Britons believe it is good at checking work for spelling, grammar and style, a YouGov survey last month showed. Others use it for more personal means: One in eight respondents say it serves as a good mental health therapist, the same proportion that believe it can act as a relationship counselor.
Yet the controversy is also useful for OpenAI. The alarm underlines an increasing reliance on AI to live our lives, further cementing OpenAI’s place in our world. The headlines, the outrage and the think pieces all reinforce one key message: ChatGPT is everywhere. It matters. The very public nature of OpenAI’s apology also furthers the sense that this technology is fundamentally on our side; there are just some kinks to iron out along the way.
I have previously reported on AI’s ability to de-indoctrinate conspiracy theorists and get them to absolve their beliefs, but the opposite is also true: ChatGPT’s positive persuasive capabilities could also, in the wrong hands, be put to manipulative ends.
Last week, an ethically dubious study conducted by Swiss researchers at the University of Zurich demonstrated the persuasive power of AI. Without informing human participants or the people controlling the online forum on the communications platform Reddit, the researchers seeded a subreddit with AI-generated comments, finding the AI was between three and six times more persuasive than humans were. (The study was approved by the university’s ethics board.) At the same time, we are being submerged under a swamp of AI-generated search results that more than half of us believe are useful, even if they fictionalize facts.
So it is worth reminding the public: AI models are not your friends. They are not designed to help you answer the questions you ask. They are designed to provide the most pleasing response possible, and to ensure that you are fully engaged with them. What happened this week was not really a bug. It was a feature.
Chris Stokel-Walker is the author of TikTok Boom: The Inside Story of the World’s Favourite App.
As the Chinese Communist Party (CCP) and its People’s Liberation Army (PLA) reach the point of confidence that they can start and win a war to destroy the democratic culture on Taiwan, any future decision to do so may likely be directly affected by the CCP’s ability to promote wars on the Korean Peninsula, in Europe, or, as most recently, on the Indian subcontinent. It stands to reason that the Trump Administration’s success early on May 10 to convince India and Pakistan to deescalate their four-day conventional military conflict, assessed to be close to a nuclear weapons exchange, also served to
The recent aerial clash between Pakistan and India offers a glimpse of how China is narrowing the gap in military airpower with the US. It is a warning not just for Washington, but for Taipei, too. Claims from both sides remain contested, but a broader picture is emerging among experts who track China’s air force and fighter jet development: Beijing’s defense systems are growing increasingly credible. Pakistan said its deployment of Chinese-manufactured J-10C fighters downed multiple Indian aircraft, although New Delhi denies this. There are caveats: Even if Islamabad’s claims are accurate, Beijing’s equipment does not offer a direct comparison
After India’s punitive precision strikes targeting what New Delhi called nine terrorist sites inside Pakistan, reactions poured in from governments around the world. The Ministry of Foreign Affairs (MOFA) issued a statement on May 10, opposing terrorism and expressing concern about the growing tensions between India and Pakistan. The statement noticeably expressed support for the Indian government’s right to maintain its national security and act against terrorists. The ministry said that it “works closely with democratic partners worldwide in staunch opposition to international terrorism” and expressed “firm support for all legitimate and necessary actions taken by the government of India
Taiwan aims to elevate its strategic position in supply chains by becoming an artificial intelligence (AI) hub for Nvidia Corp, providing everything from advanced chips and components to servers, in an attempt to edge out its closest rival in the region, South Korea. Taiwan’s importance in the AI ecosystem was clearly reflected in three major announcements Nvidia made during this year’s Computex trade show in Taipei. First, the US company’s number of partners in Taiwan would surge to 122 this year, from 34 last year, according to a slide shown during CEO Jensen Huang’s (黃仁勳) keynote speech on Monday last week.