Meta Platforms (Facebook), Alphabet (Google), Microsoft, Twitter, and a few other tech companies have come to dominate what we see and hear on the Internet, shaping billions of people’s perceptions of the world.
In pursuit of advertising revenue, their algorithms are programmed to show people content that holds attention — including extremist videos, disinformation and material designed to stimulate envy, insecurity and anger. With the rapid development of “large language models” such as ChatGPT and Bard, Big Tech’s hold on impressionable minds can only strengthen, with potentially scary consequences.
However, other outcomes are possible. Companies could deploy the latest wave of artificial intelligence much more responsibly, and two current court cases serve as warnings to those pursuing socially destructive business models.
Illustration: Constance Chou
However, public-policy interventions are also needed to break up the largest tech companies and tax digital advertising. These policy levers could help change Big Tech’s pernicious business model, thereby preventing the platforms from inflicting so much emotional harm on their users — especially vulnerable young people.
The legal cases include Gonzales v. Google, which is before the US Supreme Court. At issue is the tech industry’s insistence that Section 230 of the US Communications Decency Act exempts platform companies from any liability for third-party content that they host.
If platforms are acting more like news outlets than mere online repositories when they recommend videos, tweets or posts, they should be held to the same standard as established media, which, under existing defamation laws, are not allowed to publish what they know to be untrue.
Hence, in a US$1.6 billion lawsuit filed against Fox News, Dominion Voting Systems uncovered ample evidence that Fox’s top on-air hosts and executives were well aware (and told each other) that former US president Donald Trump’s claims of election fraud were entirely false. Dominion thus has a strong claim to damages if it can show that Fox knowingly spread falsehoods about Dominion’s voting machines in the 2020 election. Should online platforms whose algorithms disseminated the same lies not be held to the same standard?
Addressing such questions has become even more urgent now that programs such as ChatGPT are poised to reshape the Internet. These sophisticated algorithmic recommenders could potentially be trained not to promote extreme content or deliberate lies, and not to encourage extreme emotions.
If an algorithm is exploitative or manipulative toward children (or anyone else, for that matter), the responsibility for such harm should lie with the humans in charge. After all, artificial intelligence (AI) at this level is not operating autonomously of human decisionmaking. To claim otherwise is to grant their creators legal immunity.
Tech companies should no longer be able to excuse their own inattention or negligence by saying that “there is too much data” for them to monitor. That wealth of data is the source of their profit, and the sheer abundance of content on their platforms is what makes their AI so potent.
While tech firms should enjoy a reasonable degree of protection against liability for what someone else posts on their sites, this should apply only to passive content that the platforms do not in any way recommend to other users. Active content that is algorithmically pushed out to millions of people to generate revenue is a different matter. Indeed, it is just like traditional publishing, only much more powerful.
If a daily newspaper publishes a commentary by a terrorist, many readers are certain to stop subscribing, but since most individuals do not want to walk away from their existing online social networks, we need government regulation to re-empower consumers.
First, the largest platform companies should be broken up to create more intense competition between recommendation algorithms and their trainers. For this to work in the public interest, platforms also must be required to allow a user’s social network to be transferred to a different platform.
The same “interoperability” rationale allows a person to keep a mobile phone number when changing carriers. Social-media and digital content consumers should be able to vote with their feet when they do not like what a platform is promoting.
Second, and even more importantly, we need to force an adjustment in the prevailing Big Tech business model, which is based on harvesting vast amounts of user data and monetizing it through digital advertising sales. This model explains why disinformation, outrage and insecurity are so prevalent online. Emotional manipulation maximizes user engagement, enabling more intrusive data collection and higher profits.
A tax on digital advertising is one of the only practical ways to change this extraordinarily destructive business model. It would reduce platforms’ temptation to maximize user engagement through emotional manipulation. If coupled with limits on data collection, it would provide incentives to develop alternative approaches, such as subscription-based models.
Another advantage of a digital advertising tax is that it could be set even higher for content promoted to people under 21. Selling cigarettes or alcohol to minors is a serious criminal offense. While it is not feasible to forbid young people from seeing content that damages their mental health, a high rate of taxation on advertising revenues derived from promoting such material is entirely appropriate. The proceeds could be devoted to strengthening mental-health programs, not least those for teen suicide prevention.
If there is any doubt about which content is hurting young people, we can just ask the AI recommendation algorithm.
Daron Acemoglu is a professor of economics at the Massachusetts Institute of Technology (MIT). Simon Johnson is a professor at MIT’s Sloan School of Management and a former chief economist at the IMF.
Copyright: Project Syndicate
The Chinese Communist Party (CCP) continues to bully Taiwan by conducting military drills extremely close to Taiwan in late May 2024 and announcing a legal opinion in June on how they would treat “Taiwan Independence diehards” according to the PRC’s Criminal Code. This article will describe how China’s Anaconda Strategy of psychological and legal asphyxiation is employed. The CCP’s People’s Liberation Army (PLA) and Chinese Coast Guard (CCG) conducted a “punishment military exercise” against Taiwan called “Joint Sword 2024A” from 23-24 May 2024, just three days after President William Lai (賴清德) of the Democratic Progressive Party (DPP) was sworn in and
Former US president Donald Trump’s comments that Taiwan hollowed out the US semiconductor industry are incorrect. That misunderstanding could impact the future of one of the world’s most important relationships and end up aiding China at a time it is working hard to push its own tech sector to catch up. “Taiwan took our chip business from us,” the returnee US presidential contender told Bloomberg Businessweek in an interview published this week. The remarks came after the Republican nominee was asked whether he would defend Taiwan against China. It is not the first time he has said this about the nation’s
In a recent interview with the Malaysian Chinese-language newspaper Sin Chew Daily, former president Ma Ying-jeou (馬英九) called President William Lai (賴清德) “naive.” As always with Ma, one must first deconstruct what he is saying to fully understand the parallel universe he insists on defending. Who is being “naive,” Lai or Ma? The quickest way is to confront Ma with a series of pointed questions that force him to take clear stands on the complex issues involved and prevent him from his usual ramblings. Regarding China and Taiwan, the media should first begin with questions like these: “Did the Chinese Nationalist Party (KMT)
The Yomiuri Shimbun, the newspaper with the largest daily circulation in Japan, on Thursday last week published an article saying that an unidentified high-ranking Japanese official openly spoke of an analysis that the Chinese People’s Liberation Army (PLA) needs less than a week, not a month, to invade Taiwan with its amphibious forces. Reportedly, Japanese Prime Minister Fumio Kishida has already been advised of the analysis, which was based on the PLA’s military exercises last summer. A Yomiuri analysis of unclassified satellite photographs confirmed that the PLA has already begun necessary base repairs and maintenance, and is conducting amphibious operation exercises