ChatGPT, a new artificial intelligence (AI) chatbot developed by San Francisco-based research laboratory OpenAI, has taken the world by storm. Hailed as a milestone in the evolution of so-called large language models (LLMs), the world’s most famous generative AI raises important questions about who controls this nascent market and whether these powerful technologies serve the public interest.
OpenAI’s release of ChatGPT in November last year quickly became a global sensation, attracting millions of users and allegedly killing the student essay. It is able to answer questions in conversational English (along with some other languages) and perform other tasks, such as writing computer code.
The answers that ChatGPT provides are fluent and compelling, but despite its facility for language, it can sometimes make mistakes or generate factual falsehoods, a phenomenon known among AI researchers as “hallucination.”
The fear of fabricated references has led several scientific journals to ban or restrict the use of ChatGPT and similar tools in academic papers, but while the chatbot might struggle with fact-checking, it is seemingly less prone to error when it comes to programming and can easily write efficient and elegant code.
For all its flaws, ChatGPT obviously represents a major technological breakthrough, which is why Microsoft last month announced a “multiyear, multibillion-dollar investment” in OpenAI, reportedly amounting to US$10 billion, on top of US$1 billion it previously committed to the company.
Originally a nonprofit, OpenAI is now a for-profit corporation valued at US$29 billion. While it has pledged to cap its profits, its loose-fitting structure limits investors’ returns to 10,000 percent.
ChatGPT is powered by a GPT-3, a powerful LLM trained on vast amounts of text to generate natural-sounding, human-like answers. While it is the world’s most celebrated generative AI, other big tech companies such as Google and Meta have been developing their own versions. While it is still unclear how these chatbots would be monetized, a paid version of ChatGPT is reportedly forthcoming, with OpenAI projecting US$1 billion in revenues by next year.
To be sure, bad actors could abuse these tools for various illicit schemes, such as sophisticated online scams or writing malware, but the technology’s prospective applications, from coding to protein discovery, offer cause for optimism.
McKinsey estimates that 50 to 60 percent of companies have incorporated AI-powered tools, such as chatbots, into their operations.
By expanding the use of LLMs, companies could improve efficiency and productivity.
However, the massive, immensely costly and rapidly increasing computing power needed to train and maintain generative AI tools represents a substantial barrier to entry that could lead to market concentration. The potential for monopolization, together with the risk of abuse, underscores the urgent need for policymakers to consider the implications of this technological breakthrough.
Fortunately, competition authorities in the US and elsewhere seem to be aware of these risks. The British Office of Communications late last year launched an investigation of the cloud-computing market, on which all large AI models rely, while the US Federal Trade Commission is investigating Amazon Web Services (AWS), which, along with Google and Microsoft Azure, dominates the market. These investigations could have far-reaching implications for AI-powered services, which rely on enormous economies of scale.
However, it is not clear what, if anything, policymakers should do. On the one hand, if regulators do nothing, the generative AI market could end up dominated by one or two companies, like every digital market before it. On the other hand, the emergence of open-source LLMs, such as the text-to-image tool Stable Diffusion, could ensure that the market remains competitive without further intervention.
Even if for-profit models become dominant, open-source competitors could chip away at their market share, just as Mozilla’s Firefox did to Google’s Chrome browser and Android did to Apple’s mobile operating system, iOS. Then again, cloud computing giants such as AWS and Microsoft Azure could also leverage generative AI products to increase their market power.
As was debated at the World Economic Forum meeting in Davos, Switzerland, last month, generative AI is too powerful and potentially transformative to leave its fate in the hands of a few dominant companies, but while there is a clear demand for regulatory intervention, the accelerated pace of technological advance leaves governments at a huge disadvantage.
To ensure that the public interest is represented at the technological frontier, the world needs a public alternative to for-profit LLMs. Democratic governments could form a multilateral body that would develop means to prevent fakery, trolling and other online harms, like a European Centre for Nuclear Research for generative AI. Alternatively, they could establish a publicly funded competitor with a different business model and incentives to foster competition between the two models.
Whichever path global policymakers choose, standing in place is not an option. It is abundantly clear that leaving it to the market to decide how these powerful technologies are used, and by whom, is a very risky proposition.
Diane Coyle is a professor of public policy at the University of Cambridge.
Copyright: Project Syndicate
When Chinese President Xi Jinping (習近平) sits down with US President Donald Trump in Beijing on Thursday next week, Xi is unlikely to demand a dramatic public betrayal of Taiwan. He does not need to. Beijing’s preferred victory is smaller, quieter and in some ways far more dangerous: a subtle shift in American wording that appears technical, but carries major strategic meaning. The ask is simple: replace the longstanding US formulation that Washington “does not support Taiwan independence” with a harder one — that Washington “opposes” Taiwan independence. One word changes; a deterrence structure built over decades begins to shift.
Taipei is facing a severe rat infestation, and the city government is reportedly considering large-scale use of rodenticides as its primary control measure. However, this move could trigger an ecological disaster, including mass deaths of birds of prey. In the past, black kites, relatives of eagles, took more than three decades to return to the skies above the Taipei Basin. Taiwan’s black kite population was nearly wiped out by the combined effects of habitat destruction, pesticides and rodenticides. By 1992, fewer than 200 black kites remained on the island. Fortunately, thanks to more than 30 years of collective effort to preserve their remaining
After Chinese Nationalist Party (KMT) Chairwoman Cheng Li-wun (鄭麗文) met Chinese President Xi Jinping (習近平) in Beijing, most headlines referred to her as the leader of the opposition in Taiwan. Is she really, though? Being the chairwoman of the KMT does not automatically translate into being the leader of the opposition in the sense that most foreign readers would understand it. “Leader of the opposition” is a very British term. It applies to the Westminster system of parliamentary democracy, and to some extent, to other democracies. If you look at the UK right now, Conservative Party head Kemi Badenoch is
From the Iran war and nuclear weapons to tariffs and artificial intelligence, the agenda for this week’s Beijing summit between US President Donald Trump and Chinese President Xi Jinping (習近平) is packed. Xi would almost certainly bring up Taiwan, if only to demonstrate his inflexibility on the matter. However, no one needs to meet with Xi face-to-face to understand his stance. A visit to the National Museum of China in Beijing — in particular, the “Road to Rejuvenation” exhibition, which chronicles the rise and rule of the Chinese Communist Party — might be even more revealing. Xi took the members