Meta Wins in Court, but Judge Signals Future AI Copyright Battles Ahead

Meta won a major AI copyright case but the judge warned this victory won't last. He ruled the company's use of stolen books was legal while predicting future cases will go the other way—essentially giving authors a roadmap to win.

Meta Wins AI Copyright Case But Judge Warns of Future Losses

💡 TL;DR - The 30 Seconds Version

⚖️ Meta beat 13 authors in court Wednesday, but the judge warned this narrow win doesn't protect other AI companies from copyright lawsuits.

🏴‍☠️ Meta downloaded at least 666 copies of the authors' books from pirate sites after licensing talks failed, despite discussing $100 million in licensing fees.

🔬 The authors' case collapsed because Meta's AI can only reproduce a maximum of 50 words from their books, even when prompted to regurgitate content.

💰 Meta expects its AI to generate $460 billion to $1.4 trillion over the next decade, making this a high-stakes precedent for the industry.

🌊 Judge Chhabria warned AI could "flood markets" with competing content and "dramatically undermine the incentive for human beings to create."

📈 Future cases with better evidence of market harm will likely win, as the judge essentially provided a roadmap for stronger lawsuits.

Meta scored a legal victory Wednesday when a federal judge ruled that the company's use of 13 authors' books to train its AI models was legal under copyright law's "fair use" doctrine. But Judge Vince Chhabria made clear this wasn't a blank check for tech companies to raid copyrighted material.

"This ruling does not stand for the proposition that Meta's use of copyrighted materials to train its language models is lawful," Chhabria wrote. "It stands only for the proposition that these plaintiffs made the wrong arguments and failed to develop a record in support of the right one."

The case involved well-known authors including Sarah Silverman, Ta-Nehisi Coates, and Junot Díaz, who sued Meta in 2023 for using their books without permission to train its Llama AI models. Meta had downloaded the books from "shadow libraries"—essentially pirate sites—after licensing negotiations with publishers fell through.

The Authors Made the Wrong Case

The plaintiffs argued that Meta's AI could reproduce snippets of their books and that the company had harmed the market for licensing their works to AI trainers. Both arguments failed. Meta's tests showed that even when prompted to regurgitate training material, Llama could only produce a maximum of 50 words from any of the plaintiffs' books.

The licensing argument also fell flat. Courts have long held that plaintiffs can't claim harm from the loss of licensing fees for the exact use being challenged—otherwise every fair use case would automatically favor copyright holders.

But the judge saw a much stronger argument the authors largely ignored: market dilution.

The Real Threat: AI Flooding Markets

Chhabria warned that generative AI has "the potential to flood the market with endless amounts of images, songs, articles, books, and more" using "a tiny fraction of the time and creativity that would otherwise be required". "So by training generative AI models with copyrighted works, companies are creating something that often will dramatically undermine the market for those works, and thus dramatically undermine the incentive for human beings to create things the old-fashioned way."

The judge suggested that AI-generated romance novels could "successfully crowd out lesser-known works or works by up-and-coming authors." While AI books probably wouldn't affect Agatha Christie's sales, "they could very well prevent the next Agatha Christie from getting noticed or selling enough books to keep writing."

This case differed from typical copyright disputes because it "involves a technology that can generate literally millions of secondary works, with a miniscule fraction of the time and creativity used to create the original works it was trained on."

Tech Companies Shouldn't Celebrate Yet

Despite ruling for Meta, Chhabria emphasized that "in many circumstances it will be illegal to copy copyright-protected works to train generative AI models without permission". He predicted that "in cases involving uses like Meta's, it seems like the plaintiffs will often win, at least where those cases have better-developed records on the market effects of the defendant's use."

The judge also dismissed tech industry claims that copyright restrictions would kill AI development. "These products are expected to generate billions, even trillions, of dollars for the companies that are developing them," he wrote. "If using copyrighted works to train the models is as necessary as the companies say, they will figure out a way to compensate copyright holders for it."

What This Actually Means

The ruling only affects these 13 specific authors—not the "countless others whose works Meta used to train its models," since this wasn't a class action lawsuit. A separate case management conference is scheduled for July to discuss Meta's alleged distribution of copyrighted works during the downloading process.

This marks the second AI copyright victory for tech companies this week. On Monday, another San Francisco judge ruled that Anthropic's use of copyrighted books to train its Claude AI was also fair use, though the company still faces trial for allegedly pirating the books from illegal sources rather than buying them.

The Bigger Picture

The decision comes as dozens of similar lawsuits wind through courts, including cases by The New York Times against OpenAI and Disney against image generator Midjourney. The judge noted that "markets for certain types of works (like news articles) might be even more vulnerable to indirect competition from AI outputs."

Meta had initially tried to license books and discussed spending up to $100 million on licensing deals, but found that publishers generally don't hold the rights to license books for AI training—those rights belong to individual authors, and no collective licensing organization exists.

Why this matters:

  • Authors won the argument but lost the case—Chhabria essentially provided a roadmap for future plaintiffs to build stronger market dilution claims
  • The real copyright wars are just beginning—this narrow ruling suggests other cases with better evidence could easily go the other way, especially as AI models become more capable of flooding markets with competing content

❓ Frequently Asked Questions

Q: What are shadow libraries and why did Meta use them?

A: Shadow libraries are websites that offer free downloads of copyrighted books, articles, and media without permission. Meta used LibGen and Anna's Archive after publishers either ignored licensing requests or only one gave them a pricing proposal for AI training rights.

Q: How much was Meta willing to pay for book licenses?

A: Meta's head of generative AI discussed spending up to $100 million on licensing deals. However, they discovered that publishers generally don't hold AI training rights—individual authors do, and no collective licensing organization exists.

Q: How many books did Meta actually download?

A: Meta downloaded at least 666 copies of books whose copyrights the 13 plaintiffs hold. The company downloaded entire databases from LibGen and Anna's Archive, which contain millions of books and academic papers.

Q: Can Meta's AI actually reproduce the authors' books?

A: No. Even when using "adversarial prompting" designed to make AI regurgitate training data, experts could only get Llama to produce a maximum of 50 words from any plaintiff's book—and only 60% of the time.

Q: Why did the judge criticize the other AI ruling this week?

A: Judge Chhabria said Judge Alsup "focused heavily on the transformative nature of generative AI while brushing aside concerns about the harm it can inflict on the market." Chhabria believes market harm is more important than transformative use.

Q: How much money does Meta expect to make from AI?

A: Meta estimates total revenue from generative AI will range from $2-3 billion in 2025 and $460 billion to $1.4 trillion over the next ten years, despite offering Llama models for free download.

Q: What is "market dilution" and why does it matter?

A: Market dilution means AI flooding markets with competing content that hurts sales of original works. Unlike typical copyright cases involving one secondary work, AI can "generate literally millions of secondary works" using minimal time and creativity.

Q: Does this ruling protect other companies training AI on copyrighted works?

A: No. The ruling only affects these 13 authors since it wasn't a class action. The judge explicitly stated this doesn't mean Meta's use of copyrighted materials is generally lawful—just that these plaintiffs made weak arguments.

Q: What other AI copyright cases are happening right now?

A: The New York Times is suing OpenAI and Microsoft over news articles. Disney and Universal are suing Midjourney over films and TV shows. The judge noted news articles might be "even more vulnerable to indirect competition from AI outputs."

Q: Why are books especially valuable for AI training?

A: Books provide "very high-quality data" for training AI memory and handling larger amounts of text at once. They're "long but consistent," maintaining particular styles and coherent structure, plus they use proper grammar compared to internet text.

Q: What happens to Meta's separate distribution case?

A: Meta still faces a July 11 case management conference about allegedly distributing copyrighted works during the torrenting process. The judge ruled on reproduction rights but left distribution claims unresolved.

Q: Could future AI copyright cases have different outcomes?

A: Yes. The judge predicted "plaintiffs will often win" in similar cases with "better-developed records on market effects." He essentially provided a roadmap for stronger future lawsuits focusing on market dilution evidence.

Q: Why won't copyright restrictions kill AI development?

A: Companies can still license copyrighted works instead of stealing them. Since AI is expected to generate billions or trillions in revenue, companies "will figure out a way to compensate copyright holders" rather than abandon the technology entirely.

Great! You’ve successfully signed up.

Welcome back! You've successfully signed in.

You've successfully subscribed to implicator.ai.

Success! Check your email for magic link to sign-in.

Success! Your billing info has been updated.

Your billing was not updated.