Countless artists have taken inspiration from The Starry Night since Vincent van Gogh painted the swirling scene in 1889.
Now artificial intelligence (AI) systems are doing the same, training themselves on a vast collection of digitized artworks to produce new images that people can conjure in seconds from a smartphone app.
The images generated by tools such as DALL-E, Midjourney and Stable Diffusion can be weird and otherworldly, but also increasingly realistic and customizable — ask for a “peacock owl in the style of Van Gogh” and they can churn out something that might look similar to what you imagined.
Illustration: Yusha
However, while Van Gogh and other long-dead master painters are not complaining, some living artists and photographers are starting to fight back against the AI software companies creating images derived from their works.
Two new lawsuits — one from the Seattle-based photography giant Getty Images — take aim at popular image-generating services for allegedly copying and processing millions of copyright-protected images without a license.
Getty said it has begun legal proceedings in the High Court of Justice in London against Stability AI — the maker of Stable Diffusion — for infringing intellectual property rights to benefit the London-based start-up’s commercial interests.
Another lawsuit in a US federal court in San Francisco describes AI image-generators as “21st-century collage tools that violate the rights of millions of artists.”
The lawsuit, filed on Jan. 13 by three working artists on behalf of others like them, also names Stability AI as a defendant, along with San Francisco-based image-generator start-up Midjourney, and the online gallery DeviantArt.
The lawsuit alleges that AI-generated images “compete in the marketplace with the original images. Until now, when a purchaser seeks a new image ‘in the style’ of a given artist, they must pay to commission or license an original image from that artist.”
Companies that provide image-generating services typically charge users a fee. After a free trial of Midjourney through the chatting app Discord, for instance, users must buy a subscription that starts at US$10 per month or up to US$600 a year for corporate memberships. The start-up OpenAI also charges for use of its DALL-E image generator, and StabilityAI offers a paid service called DreamStudio.
“Anyone that believes that this isn’t fair use does not understand the technology and misunderstands the law,” Stability AI said in a statement.
In an interview last month, before the lawsuits were filed, Midjourney chief executive officer David Holz described his image-making service as “kind of like a search engine” pulling in a wide swath of images from across the Internet.
He compared copyright concerns about the technology with how such laws have adapted to human creativity.
“Can a person look at somebody else’s picture and learn from it and make a similar picture?” Holz said. “Obviously, it’s allowed for people and if it wasn’t, then it would destroy the whole professional art industry, probably the nonprofessional industry, too. To the extent that AIs are learning like people, it’s sort of the same thing, and if the images come out differently then it seems like it’s fine.”
The copyright disputes mark the beginning of a backlash against a new generation of impressive tools — some of them introduced just last year — that can generate new visual media, readable text and computer code on command.
They also raise broader concerns about the propensity of AI tools to amplify misinformation or cause other harm. For AI image generators, that includes the creation of nonconsensual sexual imagery.
Some systems produce photorealistic images that can be impossible to trace, making it difficult to tell the difference between what is real and what is AI. While some have safeguards in place to block offensive or harmful content, experts fear it is only a matter of time until people utilize these tools to spread disinformation and further erode public trust.
“Once we lose this capability of telling what’s real and what’s fake, everything will suddenly become fake, because you lose confidence of anything and everything,” said Wael Abd-Almageed, a professor of electrical and computer engineering at the University of Southern California.
As a test, The Associated Press submitted a text prompt on Stable Diffusion featuring the keywords “Ukraine war” and “Getty Images.” The tool created photograph-like images of soldiers in combat with warped faces and hands, pointing and carrying guns. Some of the images also featured the Getty watermark, but with garbled text.
AI can also get things wrong, such as feet and fingers, or details on ears that can sometimes give away that the images are not real, but there is no set pattern to look out for. Those visual clues can also be edited. On Midjourney, users often post on the Discord chat asking for advice on how to fix distorted faces and hands.
With some generated images traveling on social networks and potentially going viral, they can be challenging to debunk as they cannot be traced back to a specific tool or data source, said Chirag Shah, a professor at the University of Washington’s Information School, who uses the tools for research.
“You could make some guesses if you have enough experience working with these tools, but beyond that, there is no easy or scientific way to really do this,” Shah said.
For all the backlash, there are many people who embrace the new AI tools and the creativity they unleash. Some use them as a hobby to create intricate landscapes, portraits and art; others to brainstorm marketing materials, video game scenery or other ideas related to their professions.
There is plenty of room for fear, but “what else can we do with them?” asked artist Refik Anadol at the World Economic Forum in Davos, Switzerland, this month, where he displayed an exhibit of climate-themed work created by training AI models on a trove of publicly available images of coral.
At the Museum of Modern Art (MoMA) in New York, Anadol designed Unsupervised, which draws from artworks in the museum’s prestigious collection — including The Starry Night — and feeds them into a digital installation generating animations of mesmerizing colors and shapes in the museum lobby.
The installation is “constantly changing, evolving and dreaming 138,000 old artworks at MoMA’s archive,” Anadol said. “From Van Gogh to Picasso to Kandinsky, incredible, inspiring artists who defined and pioneered different techniques exist in this artwork, in this AI dream world.”
Anadol, who builds his own AI models, said in an interview that he prefers to look at the bright side of the technology, but he hopes commercial applications can be fine-tuned so that artists can more easily opt out.
“I totally hear and agree that certain artists or creators are very uncomfortable about their work being used,” he said.
For painter Erin Hanson, whose impressionist landscapes are so popular and easy to find online that she has seen their influence in AI-produced visuals, the concern is not about her own prolific output, which makes US$3 million a year. She worries about the art community as a whole.
“The original artist needs to be acknowledged in some way or compensated,” Hanson said. “That’s what copyright laws are all about, and if artists aren’t acknowledged, then it’s going to make it hard for artists to make a living in the future.”
The conflict in the Middle East has been disrupting financial markets, raising concerns about rising inflationary pressures and global economic growth. One market that some investors are particularly worried about has not been heavily covered in the news: the private credit market. Even before the joint US-Israeli attacks on Iran on Feb. 28, global capital markets had faced growing structural pressure — the deteriorating funding conditions in the private credit market. The private credit market is where companies borrow funds directly from nonbank financial institutions such as asset management companies, insurance companies and private lending platforms. Its popularity has risen since
The Donald Trump administration’s approach to China broadly, and to cross-Strait relations in particular, remains a conundrum. The 2025 US National Security Strategy prioritized the defense of Taiwan in a way that surprised some observers of the Trump administration: “Deterring a conflict over Taiwan, ideally by preserving military overmatch, is a priority.” Two months later, Taiwan went entirely unmentioned in the US National Defense Strategy, as did military overmatch vis-a-vis China, giving renewed cause for concern. How to interpret these varying statements remains an open question. In both documents, the Indo-Pacific is listed as a second priority behind homeland defense and
Every analyst watching Iran’s succession crisis is asking who would replace supreme leader Ayatollah Ali Khamenei. Yet, the real question is whether China has learned enough from the Persian Gulf to survive a war over Taiwan. Beijing purchases roughly 90 percent of Iran’s exported crude — some 1.61 million barrels per day last year — and holds a US$400 billion, 25-year cooperation agreement binding it to Tehran’s stability. However, this is not simply the story of a patron protecting an investment. China has spent years engineering a sanctions-evasion architecture that was never really about Iran — it was about Taiwan. The
In an op-ed published in Foreign Affairs on Tuesday, Chinese Nationalist Party (KMT) Chairwoman Cheng Li-wun (鄭麗文) said that Taiwan should not have to choose between aligning with Beijing or Washington, and advocated for cooperation with Beijing under the so-called “1992 consensus” as a form of “strategic ambiguity.” However, Cheng has either misunderstood the geopolitical reality and chosen appeasement, or is trying to fool an international audience with her doublespeak; nonetheless, it risks sending the wrong message to Taiwan’s democratic allies and partners. Cheng stressed that “Taiwan does not have to choose,” as while Beijing and Washington compete, Taiwan is strongest when