As the science fiction novelist William Gibson famously observed: “The future is already here — it’s just not very evenly distributed.”
I wish people would pay more attention to that adage whenever the subject of artificial intelligence (AI) comes up. Public discourse about it invariably focuses on the threat — or promise, depending on your point of view — of “superintelligent” machines, that is, ones that display human-level general intelligence, even though such devices have been 20 to 50 years away ever since we first started worrying about them.
The likelihood — or mirage — of such machines still remains a distant prospect, a point made by Andrew Ng (吳恩達), a leading AI researcher and Stanford University academic, who said that he worries about superintelligence in the same way that he frets about overpopulation on Mars.
Illustration: Lance Liu
That seems about right to me. If one were a conspiracy theorist, one might ask if our obsession with a highly speculative future has been deliberately orchestrated to divert attention from the fact that lower-level, but exceedingly powerful AI is already here and playing an ever-expanding role in shaping our economies, societies and politics.
This technology is a combination of machine learning and big data and it is everywhere, controlled and deployed by a handful of powerful corporations, with occasional walk-on parts assigned to national security agencies.
These corporations regard this version of “weak” AI as the biggest thing since sliced bread.
Google chief executive Sundar Pichai burbles about “AI everywhere” in his company’s offerings. Same goes for the other digital giants. In the face of this hype onslaught, it takes a certain amount of courage to stand up and ask awkward questions.
If this stuff is so powerful, then surely we ought to be looking at how it is being used, asking whether it is legal, ethical and good for society — and thinking about what will happen when it gets into the hands of people who are even worse than the folks who run the big tech corporations. Because it will.
Fortunately, there are academics who have started to ask these awkward questions. There are, for example, the researchers who work at AI Now, a research institute at New York University focused on the social implications of AI. Their report last year makes interesting reading.
Last week saw the publication of more in the same vein — a new critique of the technology by 26 experts from six major universities, plus a number of independent think tanks and non-governmental organizations. Its title — The Malicious Use of Artificial Intelligence: Forecasting, Prevention and Mitigation — says it all.
The report fills a serious gap in our thinking about this stuff. We have heard the hype, corporate and governmental, about the wonderful things that AI can supposedly do and we have begun to pay attention to the unintentional downsides of legitimate applications of the technology. Now the time has come to pay attention to the really malign things bad actors could do with it.
The report looks at three main “domains” in which we can expect problems. One is digital security.
The use of AI to automate tasks involved in carrying out cyberattacks will alleviate the existing trade-off between the scale and efficacy of attacks. We can also expect attacks that exploit human vulnerabilities (for example, through the use of speech synthesis for impersonation), existing software vulnerabilities (through automated hacking) or the vulnerabilities of legitimate AI systems (through corruption of the data streams on which machine learning depends).
A second threat domain is physical security — attacks with drones and autonomous weapons systems. Think v2.0 of the hobbyist drones that the Islamic State group deployed, but this time with face-recognition technology on board.
We can also expect new kinds of attacks that subvert physical systems that, for example, causes autonomous vehicles to crash, or deploy physical systems that would be impossible to remotely control from a distance, such as a 1,000-strong swarm of microdrones.
Finally, there is what the authors call “political security” — using AI to automate tasks involved in surveillance, persuasion (creating targeted propaganda) and deception (such as, manipulating videos).
We can also expect new kinds of attacks based on machine-learning’s capability to infer human behaviors, moods and beliefs from available data. This technology will obviously be welcomed by authoritarian states, but it will also further undermine the ability of democracies to sustain truthful public debates.
The bots and fake Facebook accounts that currently pollute our public sphere will look awfully amateurish in a couple of years.
The report is available as a free download and is worth reading in full.
If it were about the dangers of future or speculative technologies, then it might be reasonable to dismiss it as academic scaremongering. The alarming thing is that most of the problematic capabilities that its authors envisage are already available and in many cases are embedded in many of the networked services that we use every day.
Gibson was right: The future has already arrived.
As the war in Burma stretches into its 76th year, China continues to play both sides. Beijing backs the junta, which seized power in the 2021 coup, while also funding some of the resistance groups fighting the regime. Some suggest that Chinese President Xi Jinping (習近平) is hedging his bets, positioning China to side with the victors regardless of the outcome. However, a more accurate explanation is that China is acting pragmatically to safeguard its investments and ensure the steady flow of natural resources and energy for its economy. China’s primary interest is stability and supporting the junta initially seemed like the best
The US election result will significantly impact its foreign policy with global implications. As tensions escalate in the Taiwan Strait and conflicts elsewhere draw attention away from the western Pacific, Taiwan was closely monitoring the election, as many believe that whoever won would confront an increasingly assertive China, especially with speculation over a potential escalation in or around 2027. A second Donald Trump presidency naturally raises questions concerning the future of US policy toward China and Taiwan, with Trump displaying mixed signals as to his position on the cross-strait conflict. US foreign policy would also depend on Trump’s Cabinet and
Navy Commander Admiral Tang Hua (唐華) said in an interview with The Economist that the Chinese People’s Liberation Army (PLA) has been implementing an “anaconda strategy” to subdue Taiwan since President William Lai (賴清德) assumed office. The Chinese military is “slowly, but surely” increasing its presence around Taiwan proper, it quoted Tang as saying. “They are ready to blockade Taiwan at any time they want,” he said. “They give you extreme pressure, pressure, pressure. They’re trying to exhaust you.” Beijing’s goal is to “force Taiwan to make mistakes,” Tang said, adding that they could be “excuses” for a blockade. The interview reminds me
Republican candidate and former US president Donald Trump is to be the 47th president of the US after beating his Democratic rival, US Vice President Kamala Harris, in the election on Tuesday. Trump’s thumping victory — winning 295 Electoral College votes against Harris’ 226 as of press time last night, along with the Republicans winning control of the US Senate and possibly the House of Representatives — is a remarkable political comeback from his 2020 defeat to US President Joe Biden, and means Trump has a strong political mandate to implement his agenda. What does Trump’s victory mean for Taiwan, Asia, deterrence