Imagine receiving a telephone call from your aging mother seeking your help because she has forgotten her banking password.
Except it is not your mother. The voice on the other end of the phone call just sounds deceptively like her.
It is actually a synthesized voice, a tour de force of artificial intelligence (AI) technology that has been crafted to make it possible for someone to masquerade via the telephone.
Illustration: Mountain People
Such a situation is still science fiction — but just barely. It is also the future of crime.
The software components necessary to make such masking technology widely accessible are advancing rapidly.
Recently, for example, DeepMind, the Alphabet subsidiary known for a program that has bested some of the top human players in the 3,000-year-old Chinese board game go, announced that it had designed a program that “mimics any human voice and which sounds more natural than the best existing text-to-speech systems, reducing the gap with human performance by over 50 percent.”
EXPONENTIAL SCALE
The irony, of course, is that this year the computer security industry, with US$75 billion in annual revenue, has started to talk about how machine learning and pattern recognition techniques will improve the woeful state of computer security.
However, there is a downside.
“The thing people don’t get is that cybercrime is becoming automated and it is scaling exponentially,” said Marc Goodman, a law enforcement agency adviser and the author of Future Crimes.
“This is not about Matthew Broderick hacking from his basement,” he added, referring to the 1983 movie War Games.
The alarm about malevolent use of advanced AI technologies was sounded earlier this year by US Director of National Intelligence James Clapper. In his annual review of security, Clapper underscored the point that while AI systems would make some things easier, they would also expand the vulnerabilities of the online world.
The growing sophistication of computer criminals can be seen in the evolution of attack tools like the widely used malicious program known as Blackshades, Goodman said.
The author of the program, a Swedish national, was convicted last year in the US.
The system, which was sold widely in the computer underground, functioned as a “criminal franchise in a box,” Goodman said.
It allowed users without technical skills to deploy computer ransomware or perform video or audio eavesdropping with a mouse click.
The next generation of these tools will add machine learning capabilities that have been pioneered by AI researchers to improve the quality of machine vision, speech understanding, speech synthesis and natural language understanding.
Some computer security researchers believe that digital criminals have been experimenting with the use of AI technologies for more than half a decade.
That can be seen in efforts to subvert the Internet’s omnipresent CAPTCHA — Completely Automated Public Turing test to tell Computers and Humans Apart — the challenge and response puzzle invented in 2003 by Carnegie Mellon University researchers to block automated programs from stealing online accounts.
Both “white hat” AI researchers and “black hat” criminals have been deploying machine vision software to subvert CAPTCHAs for more than half a decade, said Stefan Savage, a computer security researcher at the University of California, San Diego.
“If you don’t change your CAPTCHA for two years, you will be owned by some machine vision algorithm,” he said.
Surprisingly, one thing that has slowed the development of malicious AI has been the ready availability of either low-cost or free human labor. For example, some cybercriminals have farmed out CAPTCHA-breaking schemes to electronic sweatshops where humans are used to decode the puzzles for a tiny fee.
Even more inventive computer crooks have used online pornography as a reward for human Web surfers who break the CAPTCHA, Goodman said.
Free labor is a commodity that AI software will not be able to compete with any time soon.
So what is next?
SOCIAL ENGINEERING
Criminals, for starters, can piggyback on new technology developments. Voice-recognition technology like Apple’s Siri and Microsoft’s Cortana are now used extensively to interact with computers. And Amazon’s Echo voice-controlled speaker and Facebook’s Messenger chatbot platform are rapidly becoming conduits for online commerce and customer support.
As is often the case, whenever a communication advancement like voice recognition starts to go mainstream, criminals looking to take advantage of it are not far behind.
“I would argue that companies that offer customer support via chatbots are unwittingly making themselves liable to social engineering,” said Brian Krebs, an investigative reporter who publishes at krebsonsecurity.com.
Social engineering, which refers to the practice of manipulating people into performing actions or divulging information, is widely seen as the weakest link in the computer security chain.
Cybercriminals already exploit the best qualities in humans — trust and willingness to help others — to steal and spy. The ability to create AI avatars that can fool people online will only make the problem worse.
This can already be seen in efforts by US state governments and political campaigns to use chatbot technology for political propaganda.
Researchers have coined the term “computational propaganda” to describe the explosion of deceptive social media campaigns on services like Facebook and Twitter.
In a recent research paper, Philip Howard, a sociologist at the Oxford Internet Institute, and Bence Kollanyi, a researcher at Corvinus University of Budapest, described how political chatbots had a “small but strategic role” in shaping the online conversation during the run-up to the “Brexit” referendum.
It is only a matter of time before such software is put to criminal use.
“There’s a lot of cleverness in designing social engineering attacks, but as far as I know, nobody has yet started using machine learning to find the highest quality suckers,” said Mark Seiden, an independent computer security specialist.
He paused and added, “I should have replied: ‘I’m sorry, Dave, I can’t answer that question right now.’”
The gutting of Voice of America (VOA) and Radio Free Asia (RFA) by US President Donald Trump’s administration poses a serious threat to the global voice of freedom, particularly for those living under authoritarian regimes such as China. The US — hailed as the model of liberal democracy — has the moral responsibility to uphold the values it champions. In undermining these institutions, the US risks diminishing its “soft power,” a pivotal pillar of its global influence. VOA Tibetan and RFA Tibetan played an enormous role in promoting the strong image of the US in and outside Tibet. On VOA Tibetan,
Former minister of culture Lung Ying-tai (龍應台) has long wielded influence through the power of words. Her articles once served as a moral compass for a society in transition. However, as her April 1 guest article in the New York Times, “The Clock Is Ticking for Taiwan,” makes all too clear, even celebrated prose can mislead when romanticism clouds political judgement. Lung crafts a narrative that is less an analysis of Taiwan’s geopolitical reality than an exercise in wistful nostalgia. As political scientists and international relations academics, we believe it is crucial to correct the misconceptions embedded in her article,
Sung Chien-liang (宋建樑), the leader of the Chinese Nationalist Party’s (KMT) efforts to recall Democratic Progressive Party (DPP) Legislator Lee Kun-cheng (李坤城), caused a national outrage and drew diplomatic condemnation on Tuesday after he arrived at the New Taipei City District Prosecutors’ Office dressed in a Nazi uniform. Sung performed a Nazi salute and carried a copy of Adolf Hitler’s Mein Kampf as he arrived to be questioned over allegations of signature forgery in the recall petition. The KMT’s response to the incident has shown a striking lack of contrition and decency. Rather than apologizing and distancing itself from Sung’s actions,
US President Trump weighed into the state of America’s semiconductor manufacturing when he declared, “They [Taiwan] stole it from us. They took it from us, and I don’t blame them. I give them credit.” At a prior White House event President Trump hosted TSMC chairman C.C. Wei (魏哲家), head of the world’s largest and most advanced chip manufacturer, to announce a commitment to invest US$100 billion in America. The president then shifted his previously critical rhetoric on Taiwan and put off tariffs on its chips. Now we learn that the Trump Administration is conducting a “trade investigation” on semiconductors which