“Ordinary people here in China aren’t happy about this technology, but they have no choice. If the police say there have to be cameras in a community, people will just have to live with it. There’s always that demand and we’re here to fulfill it,” said Chen Wei (陳偉), founder and chairman of Taigusys, a company specializing in emotion recognition technology, the latest evolution in the broader world of surveillance systems that play a part in nearly every aspect of Chinese society.
Emotion-recognition technologies — in which facial expressions of anger, sadness, happiness and boredom as well as other biometric data are tracked — are supposedly able to infer a person’s feelings based on traits such as facial muscle movements, vocal tone, body movements and other biometric signals.
It goes beyond facial-recognition technologies, which simply compare faces to determine a match, but similar to it, the new technology involves the mass collection of sensitive personal data to track, monitor and profile people, and uses machine learning to analyze expressions and other clues.
Illustration: Constance Chou
The industry is booming in China, where since at least 2012, figures including President Xi Jinping (習近平) have emphasized the creation of “positive energy” as part of an ideological campaign to encourage certain kinds of expression and limit others.
Critics say that the technology is based on a pseudo-science of stereotypes, and an increasing number of researchers, lawyers and civil rights advocates believe that it has serious implications for human rights, privacy and freedom of expression.
With the global industry forecast to be worth nearly US$36 billion by 2023, growing nearly 30 percent annually, rights groups say that action needs to be taken now.
The main office of Taigusys is tucked behind a few low-rise office buildings in the southern Chinese boomtown of Shenzhen. Visitors are greeted at the door by a series of cameras capturing their images on a big screen that displays body temperature, along with age estimates and other statistics.
Chen says that the system in the doorway is the company’s best seller because of high demand during the COVID-19 pandemic.
Chen hails emotion recognition as a way to predict dangerous behavior by prisoners, detect potential criminals at police checkpoints, problem pupils in schools and older people with dementia in care homes.
Taigusys systems are installed in about 300 prisons, detention centers and remand facilities across China, connecting 60,000 cameras.
“Violence and suicide are very common in detention centers,” Chen said. “Even if police nowadays don’t beat prisoners, they often try to wear them down by not allowing them to fall asleep. As a result, some prisoners will have a mental breakdown and seek to kill themselves, and our system will help prevent that from happening.”
Since prisoners know that they are monitored by the system — 24 hours a day, in real time — they are made more docile, which for authorities is a positive on many fronts, Chen said.
“Because they know what the system does, they won’t consciously try to violate certain rules,” he said.
Besides prisons and police checkpoints, Taigusys has deployed its systems in schools to monitor teachers, pupils and staff, in care homes for older people to detect falls and changes in the emotional state of residents, and in shopping centers and car parks.
While the use of emotion-recognition technology in schools has sparked some criticism, there has been very little discussion of its use by authorities on citizens.
Chen, while aware of the concerns, played up the system’s potential to stop violent incidents. He cited an incident where a security guard in June last year stabbed 41 people in Guangxi Province in southern China, saying that it was technologically preventable.
Vidushi Marda is a digital program manager at the British human rights organization Article 19 and a lawyer focused on the socio-legal implications of emerging technologies. She disputes Chen’s view on the Guangxi stabbing.
“This is a familiar and slightly frustrating narrative that we see used frequently when newer, shiny technologies are introduced under the umbrella of safety or security, but in reality, video surveillance has little nexus to safety, and I’m not sure how they thought that feedback in real time would fix violence,” Marda said.
“A lot of biometric surveillance, I think, is closely tied to intimidation and censorship, and I suppose [emotion recognition] is one example of just that,” she said.
An Article 19 report on the development of these surveillance technologies — which one Chinese firm describes as “biometrics 3.0” — by 27 companies in China found that its growth without safeguards and public deliberation was especially problematic, particularly in the public security and education sectors.
Ultimately, groups such as Article 19 say that the technology should be banned before widespread adoption globally makes the ramifications too difficult to contain.
The Guardian contacted a range of companies covered in the report. Only Taigusys responded to an interview request.
Another problem is that recognition systems are usually based on actors posing in what they think are happy, sad, angry and other emotional states, and not on real expressions of those emotions. Facial expressions can also vary widely across cultures, leading to inaccuracies and ethnic bias.
One Taigusys system that is used by police in China, as well as security services in Thailand and some African countries, includes identifiers such as “yellow, white, black” and even “Uighur.”
“The populations in these countries are more racially diverse than in China, and in China, it’s also used to tell Uighurs from Han Chinese,” Chen said, referring to the country’s dominant ethnicity. “If an Uighur appears, they will be tagged, but it won’t tag Han Chinese.”
Asked if he was concerned about these features being misused by authorities, Chen said that he is not worried because the software is used by police, implying that such institutions should automatically be trusted.
“I’m not concerned because it’s not our technology that’s the problem,” Chen said. “There are demands for this technology in certain scenarios and places, and we will try our best to meet those demands.”
For Shazeda Ahmed, an artificial intelligence researcher at New York University who contributed to the Article 19 report, these are all “terrible reasons.”
“That Chinese conceptions of race are going to be built into technology and exported to other parts of the world is really troubling, particularly since there isn’t the kind of critical discourse [about racism and ethnicity in China] that we’re having in the United States,” she said.
“If anything, research and investigative reporting over the last few years have shown that sensitive personal information is particularly dangerous when in the hands of state entities, especially given the wide ambit of their possible use by state actors,” Ahmed said.
One driver of the emotion-recognition technology sector in China is the country’s lack of strict privacy laws. There are essentially no laws restricting the authorities’ access to biometric data on grounds of national security or public safety, which gives companies such as Taigusys complete freedom to develop and roll out these products when similar businesses in the US, Japan or Europe cannot, Chen said.
“So we have the chance to gather as much information as possible and find the best scenarios to make use of that data,” he said.
Xiaomi Corp founder Lei Jun (雷軍) on May 22 made a high-profile announcement, giving online viewers a sneak peek at the company’s first 3-nanometer mobile processor — the Xring O1 chip — and saying it is a breakthrough in China’s chip design history. Although Xiaomi might be capable of designing chips, it lacks the ability to manufacture them. No matter how beautifully planned the blueprints are, if they cannot be mass-produced, they are nothing more than drawings on paper. The truth is that China’s chipmaking efforts are still heavily reliant on the free world — particularly on Taiwan Semiconductor Manufacturing
Keelung Mayor George Hsieh (謝國樑) of the Chinese Nationalist Party (KMT) on Tuesday last week apologized over allegations that the former director of the city’s Civil Affairs Department had illegally accessed citizens’ data to assist the KMT in its campaign to recall Democratic Progressive Party (DPP) councilors. Given the public discontent with opposition lawmakers’ disruptive behavior in the legislature, passage of unconstitutional legislation and slashing of the central government’s budget, civic groups have launched a massive campaign to recall KMT lawmakers. The KMT has tried to fight back by initiating campaigns to recall DPP lawmakers, but the petition documents they
A recent scandal involving a high-school student from a private school in Taichung has reignited long-standing frustrations with Taiwan’s increasingly complex and high-pressure university admissions system. The student, who had successfully gained admission to several prestigious medical schools, shared their learning portfolio on social media — only for Internet sleuths to quickly uncover a falsified claim of receiving a “Best Debater” award. The fallout was swift and unforgiving. National Yang Ming Chiao Tung University and Taipei Medical University revoked the student’s admission on Wednesday. One day later, Chung Shan Medical University also announced it would cancel the student’s admission. China Medical
Construction of the Ma-anshan Nuclear Power Plant in Pingtung County’s Hengchun Township (恆春) started in 1978. It began commercial operations in 1984. Since then, it has experienced several accidents, radiation pollution and fires. It was finally decommissioned on May 17 after the operating license of its No. 2 reactor expired. However, a proposed referendum to be held on Aug. 23 on restarting the reactor is potentially bringing back those risks. Four reasons are listed for holding the referendum: First, the difficulty of meeting greenhouse gas reduction targets and the inefficiency of new energy sources such as photovoltaic and wind power. Second,