Countless artists have taken inspiration from The Starry Night since Vincent van Gogh painted the swirling scene in 1889.
Now artificial intelligence (AI) systems are doing the same, training themselves on a vast collection of digitized artworks to produce new images that people can conjure in seconds from a smartphone app.
The images generated by tools such as DALL-E, Midjourney and Stable Diffusion can be weird and otherworldly, but also increasingly realistic and customizable — ask for a “peacock owl in the style of Van Gogh” and they can churn out something that might look similar to what you imagined.
Illustration: Yusha
However, while Van Gogh and other long-dead master painters are not complaining, some living artists and photographers are starting to fight back against the AI software companies creating images derived from their works.
Two new lawsuits — one from the Seattle-based photography giant Getty Images — take aim at popular image-generating services for allegedly copying and processing millions of copyright-protected images without a license.
Getty said it has begun legal proceedings in the High Court of Justice in London against Stability AI — the maker of Stable Diffusion — for infringing intellectual property rights to benefit the London-based start-up’s commercial interests.
Another lawsuit in a US federal court in San Francisco describes AI image-generators as “21st-century collage tools that violate the rights of millions of artists.”
The lawsuit, filed on Jan. 13 by three working artists on behalf of others like them, also names Stability AI as a defendant, along with San Francisco-based image-generator start-up Midjourney, and the online gallery DeviantArt.
The lawsuit alleges that AI-generated images “compete in the marketplace with the original images. Until now, when a purchaser seeks a new image ‘in the style’ of a given artist, they must pay to commission or license an original image from that artist.”
Companies that provide image-generating services typically charge users a fee. After a free trial of Midjourney through the chatting app Discord, for instance, users must buy a subscription that starts at US$10 per month or up to US$600 a year for corporate memberships. The start-up OpenAI also charges for use of its DALL-E image generator, and StabilityAI offers a paid service called DreamStudio.
“Anyone that believes that this isn’t fair use does not understand the technology and misunderstands the law,” Stability AI said in a statement.
In an interview last month, before the lawsuits were filed, Midjourney chief executive officer David Holz described his image-making service as “kind of like a search engine” pulling in a wide swath of images from across the Internet.
He compared copyright concerns about the technology with how such laws have adapted to human creativity.
“Can a person look at somebody else’s picture and learn from it and make a similar picture?” Holz said. “Obviously, it’s allowed for people and if it wasn’t, then it would destroy the whole professional art industry, probably the nonprofessional industry, too. To the extent that AIs are learning like people, it’s sort of the same thing, and if the images come out differently then it seems like it’s fine.”
The copyright disputes mark the beginning of a backlash against a new generation of impressive tools — some of them introduced just last year — that can generate new visual media, readable text and computer code on command.
They also raise broader concerns about the propensity of AI tools to amplify misinformation or cause other harm. For AI image generators, that includes the creation of nonconsensual sexual imagery.
Some systems produce photorealistic images that can be impossible to trace, making it difficult to tell the difference between what is real and what is AI. While some have safeguards in place to block offensive or harmful content, experts fear it is only a matter of time until people utilize these tools to spread disinformation and further erode public trust.
“Once we lose this capability of telling what’s real and what’s fake, everything will suddenly become fake, because you lose confidence of anything and everything,” said Wael Abd-Almageed, a professor of electrical and computer engineering at the University of Southern California.
As a test, The Associated Press submitted a text prompt on Stable Diffusion featuring the keywords “Ukraine war” and “Getty Images.” The tool created photograph-like images of soldiers in combat with warped faces and hands, pointing and carrying guns. Some of the images also featured the Getty watermark, but with garbled text.
AI can also get things wrong, such as feet and fingers, or details on ears that can sometimes give away that the images are not real, but there is no set pattern to look out for. Those visual clues can also be edited. On Midjourney, users often post on the Discord chat asking for advice on how to fix distorted faces and hands.
With some generated images traveling on social networks and potentially going viral, they can be challenging to debunk as they cannot be traced back to a specific tool or data source, said Chirag Shah, a professor at the University of Washington’s Information School, who uses the tools for research.
“You could make some guesses if you have enough experience working with these tools, but beyond that, there is no easy or scientific way to really do this,” Shah said.
For all the backlash, there are many people who embrace the new AI tools and the creativity they unleash. Some use them as a hobby to create intricate landscapes, portraits and art; others to brainstorm marketing materials, video game scenery or other ideas related to their professions.
There is plenty of room for fear, but “what else can we do with them?” asked artist Refik Anadol at the World Economic Forum in Davos, Switzerland, this month, where he displayed an exhibit of climate-themed work created by training AI models on a trove of publicly available images of coral.
At the Museum of Modern Art (MoMA) in New York, Anadol designed Unsupervised, which draws from artworks in the museum’s prestigious collection — including The Starry Night — and feeds them into a digital installation generating animations of mesmerizing colors and shapes in the museum lobby.
The installation is “constantly changing, evolving and dreaming 138,000 old artworks at MoMA’s archive,” Anadol said. “From Van Gogh to Picasso to Kandinsky, incredible, inspiring artists who defined and pioneered different techniques exist in this artwork, in this AI dream world.”
Anadol, who builds his own AI models, said in an interview that he prefers to look at the bright side of the technology, but he hopes commercial applications can be fine-tuned so that artists can more easily opt out.
“I totally hear and agree that certain artists or creators are very uncomfortable about their work being used,” he said.
For painter Erin Hanson, whose impressionist landscapes are so popular and easy to find online that she has seen their influence in AI-produced visuals, the concern is not about her own prolific output, which makes US$3 million a year. She worries about the art community as a whole.
“The original artist needs to be acknowledged in some way or compensated,” Hanson said. “That’s what copyright laws are all about, and if artists aren’t acknowledged, then it’s going to make it hard for artists to make a living in the future.”
On May 7, 1971, Henry Kissinger planned his first, ultra-secret mission to China and pondered whether it would be better to meet his Chinese interlocutors “in Pakistan where the Pakistanis would tape the meeting — or in China where the Chinese would do the taping.” After a flicker of thought, he decided to have the Chinese do all the tape recording, translating and transcribing. Fortuitously, historians have several thousand pages of verbatim texts of Dr. Kissinger’s negotiations with his Chinese counterparts. Paradoxically, behind the scenes, Chinese stenographers prepared verbatim English language typescripts faster than they could translate and type them
More than 30 years ago when I immigrated to the US, applied for citizenship and took the 100-question civics test, the one part of the naturalization process that left the deepest impression on me was one question on the N-400 form, which asked: “Have you ever been a member of, involved in or in any way associated with any communist or totalitarian party anywhere in the world?” Answering “yes” could lead to the rejection of your application. Some people might try their luck and lie, but if exposed, the consequences could be much worse — a person could be fined,
Xiaomi Corp founder Lei Jun (雷軍) on May 22 made a high-profile announcement, giving online viewers a sneak peek at the company’s first 3-nanometer mobile processor — the Xring O1 chip — and saying it is a breakthrough in China’s chip design history. Although Xiaomi might be capable of designing chips, it lacks the ability to manufacture them. No matter how beautifully planned the blueprints are, if they cannot be mass-produced, they are nothing more than drawings on paper. The truth is that China’s chipmaking efforts are still heavily reliant on the free world — particularly on Taiwan Semiconductor Manufacturing
Keelung Mayor George Hsieh (謝國樑) of the Chinese Nationalist Party (KMT) on Tuesday last week apologized over allegations that the former director of the city’s Civil Affairs Department had illegally accessed citizens’ data to assist the KMT in its campaign to recall Democratic Progressive Party (DPP) councilors. Given the public discontent with opposition lawmakers’ disruptive behavior in the legislature, passage of unconstitutional legislation and slashing of the central government’s budget, civic groups have launched a massive campaign to recall KMT lawmakers. The KMT has tried to fight back by initiating campaigns to recall DPP lawmakers, but the petition documents they