ChatGPT, a new artificial intelligence (AI) chatbot developed by San Francisco-based research laboratory OpenAI, has taken the world by storm. Hailed as a milestone in the evolution of so-called large language models (LLMs), the world’s most famous generative AI raises important questions about who controls this nascent market and whether these powerful technologies serve the public interest.
OpenAI’s release of ChatGPT in November last year quickly became a global sensation, attracting millions of users and allegedly killing the student essay. It is able to answer questions in conversational English (along with some other languages) and perform other tasks, such as writing computer code.
The answers that ChatGPT provides are fluent and compelling, but despite its facility for language, it can sometimes make mistakes or generate factual falsehoods, a phenomenon known among AI researchers as “hallucination.”
The fear of fabricated references has led several scientific journals to ban or restrict the use of ChatGPT and similar tools in academic papers, but while the chatbot might struggle with fact-checking, it is seemingly less prone to error when it comes to programming and can easily write efficient and elegant code.
For all its flaws, ChatGPT obviously represents a major technological breakthrough, which is why Microsoft last month announced a “multiyear, multibillion-dollar investment” in OpenAI, reportedly amounting to US$10 billion, on top of US$1 billion it previously committed to the company.
Originally a nonprofit, OpenAI is now a for-profit corporation valued at US$29 billion. While it has pledged to cap its profits, its loose-fitting structure limits investors’ returns to 10,000 percent.
ChatGPT is powered by a GPT-3, a powerful LLM trained on vast amounts of text to generate natural-sounding, human-like answers. While it is the world’s most celebrated generative AI, other big tech companies such as Google and Meta have been developing their own versions. While it is still unclear how these chatbots would be monetized, a paid version of ChatGPT is reportedly forthcoming, with OpenAI projecting US$1 billion in revenues by next year.
To be sure, bad actors could abuse these tools for various illicit schemes, such as sophisticated online scams or writing malware, but the technology’s prospective applications, from coding to protein discovery, offer cause for optimism.
McKinsey estimates that 50 to 60 percent of companies have incorporated AI-powered tools, such as chatbots, into their operations.
By expanding the use of LLMs, companies could improve efficiency and productivity.
However, the massive, immensely costly and rapidly increasing computing power needed to train and maintain generative AI tools represents a substantial barrier to entry that could lead to market concentration. The potential for monopolization, together with the risk of abuse, underscores the urgent need for policymakers to consider the implications of this technological breakthrough.
Fortunately, competition authorities in the US and elsewhere seem to be aware of these risks. The British Office of Communications late last year launched an investigation of the cloud-computing market, on which all large AI models rely, while the US Federal Trade Commission is investigating Amazon Web Services (AWS), which, along with Google and Microsoft Azure, dominates the market. These investigations could have far-reaching implications for AI-powered services, which rely on enormous economies of scale.
However, it is not clear what, if anything, policymakers should do. On the one hand, if regulators do nothing, the generative AI market could end up dominated by one or two companies, like every digital market before it. On the other hand, the emergence of open-source LLMs, such as the text-to-image tool Stable Diffusion, could ensure that the market remains competitive without further intervention.
Even if for-profit models become dominant, open-source competitors could chip away at their market share, just as Mozilla’s Firefox did to Google’s Chrome browser and Android did to Apple’s mobile operating system, iOS. Then again, cloud computing giants such as AWS and Microsoft Azure could also leverage generative AI products to increase their market power.
As was debated at the World Economic Forum meeting in Davos, Switzerland, last month, generative AI is too powerful and potentially transformative to leave its fate in the hands of a few dominant companies, but while there is a clear demand for regulatory intervention, the accelerated pace of technological advance leaves governments at a huge disadvantage.
To ensure that the public interest is represented at the technological frontier, the world needs a public alternative to for-profit LLMs. Democratic governments could form a multilateral body that would develop means to prevent fakery, trolling and other online harms, like a European Centre for Nuclear Research for generative AI. Alternatively, they could establish a publicly funded competitor with a different business model and incentives to foster competition between the two models.
Whichever path global policymakers choose, standing in place is not an option. It is abundantly clear that leaving it to the market to decide how these powerful technologies are used, and by whom, is a very risky proposition.
Diane Coyle is a professor of public policy at the University of Cambridge.
Copyright: Project Syndicate
As strategic tensions escalate across the vast Indo-Pacific region, Taiwan has emerged as more than a potential flashpoint. It is the fulcrum upon which the credibility of the evolving American-led strategy of integrated deterrence now rests. How the US and regional powers like Japan respond to Taiwan’s defense, and how credible the deterrent against Chinese aggression proves to be, will profoundly shape the Indo-Pacific security architecture for years to come. A successful defense of Taiwan through strengthened deterrence in the Indo-Pacific would enhance the credibility of the US-led alliance system and underpin America’s global preeminence, while a failure of integrated deterrence would
It is being said every second day: The ongoing recall campaign in Taiwan — where citizens are trying to collect enough signatures to trigger re-elections for a number of Chinese Nationalist Party (KMT) legislators — is orchestrated by the Democratic Progressive Party (DPP), or even President William Lai (賴清德) himself. The KMT makes the claim, and foreign media and analysts repeat it. However, they never show any proof — because there is not any. It is alarming how easily academics, journalists and experts toss around claims that amount to accusing a democratic government of conspiracy — without a shred of evidence. These
The Executive Yuan recently revised a page of its Web site on ethnic groups in Taiwan, replacing the term “Han” (漢族) with “the rest of the population.” The page, which was updated on March 24, describes the composition of Taiwan’s registered households as indigenous (2.5 percent), foreign origin (1.2 percent) and the rest of the population (96.2 percent). The change was picked up by a social media user and amplified by local media, sparking heated discussion over the weekend. The pan-blue and pro-China camp called it a politically motivated desinicization attempt to obscure the Han Chinese ethnicity of most Taiwanese.
On Wednesday last week, the Rossiyskaya Gazeta published an article by Chinese President Xi Jinping (習近平) asserting the People’s Republic of China’s (PRC) territorial claim over Taiwan effective 1945, predicated upon instruments such as the 1943 Cairo Declaration and the 1945 Potsdam Proclamation. The article further contended that this de jure and de facto status was subsequently reaffirmed by UN General Assembly Resolution 2758 of 1971. The Ministry of Foreign Affairs promptly issued a statement categorically repudiating these assertions. In addition to the reasons put forward by the ministry, I believe that China’s assertions are open to questions in international