The artificial intelligence industry felt a seismic moment earlier this month when Anthropic PBC agreed to pay $1.5 billion to settle the largest copyright lawsuit in US history.
While Judge William Alsup initially denied the proposed agreement as “nowhere close to complete,” the landmark deal—involving roughly 500,000 works that were allegedly pirated for AI training—offers crucial insights into how the tech sector must navigate intellectual property rights.
Why It Matters
The proposed Anthropic settlement is more than a monumental payout—it sets a practical framework for how AI companies must approach training-data acquisition.
Alsup had ruled in June that using copyrighted books to train AI models constitutes transformative fair use but said that downloading pirated copies from shadow libraries violates copyright law. The ruling drew a nuanced distinction: How AI companies acquire data matters as much as what they do with it, if not more.
Although Alsup’s initial rejection focused on procedural concerns, including undefined claim processes, the underlying settlement structure remains instructive. With similar copyright lawsuits pending nationwide against AI developers, including Disney Enterprises v. Midjourney Inc. and Kadrey v. Meta Platforms Inc., this framework plots a potential roadmap for resolution.
Settlement Structure
Under the proposed agreement’s principal terms, Anthropic would pay about $3,000 per work to the authors, destroy all copies of works obtained from pirate libraries, and crucially receive only a limited release covering past training behavior.
The company would gain no license for future training and thus remain exposed to potential lawsuits over AI-generated outputs that might infringe existing works.
This structured approach reflects sophisticated legal thinking. By limiting the release to historical conduct, the settlement would preserve authors’ rights while allowing Anthropic to move forward with a clean slate—if it legitimately sources training data.
Economic Reality
The proposed settlement reveals the harsh economic calculus facing AI developers who rely on shady data sources.
Anthropic could have faced statutory damages up to $150,000 per work, potentially exceeding $1 trillion in total exposure, so the $3,000-per-work settlement represents a massive discount. But it would come at the cost of extensive litigation and existential business risk.
AI companies still in development or scaling phases should invest in legitimate data acquisition upfront. The cost of negotiating licensing agreements or buying content legally pales in comparison with the litigation exposure and reputational damage of using pirated or otherwise illicit training materials.
Companies that proactively establish legitimate data pipelines will avoid Anthropic’s predicament entirely.
Licensing Evolution
Perhaps most significantly, the settlement—if approved—could accelerate development of a mature licensing market for AI-training data.
Industry observers are already drawing parallels to how the music industry evolved after Napster, ultimately creating sustainable licensing frameworks that compensated rights holders while enabling technological innovation.
We’ll likely see the emergence of collective licensing organizations, similar to performance rights organizations in the music industry, that can aggregate content efficiently and negotiate blanket licenses for AI training.
This development would help both sides, as AI companies gain legal certainty and scalable access to high-quality content while creators receive compensation and maintain control over their works.
Practical Takeaways
For attorneys advising AI companies, several considerations emerge from both the proposed settlement and its initial rejection:
Due diligence is paramount. Document all data sources meticulously and maintain clear chain-of-custody records. The court’s focus on how data was acquired means that procurement methods will face intense scrutiny in future litigation.
Fair use remains viable but narrow. Alsup’s ruling confirms that AI training can qualify as transformative fair use, but only when the underlying materials are lawfully obtained. This creates a safe harbor for companies that invest in proper licensing.
Class action exposure is real. The potential for massive statutory damages poses an existential risk to companies that rely on questionable data sources.
Looking Ahead
Whether or not this specific settlement gains approval, it fundamentally signals how the landscape is changing.
Caution should come to companies intending to hide behind fair use while using pirated content. The path forward requires investment in legitimate data acquisition, proactive licensing negotiations, and careful documentation of all training materials.
For an industry built on learning from existing creative content, that shouldn’t be viewed as a limitation. It should be the foundation for sustainable innovation that helps both creators and technologists.
The case is Bartz v. Anthropic PBC, N.D. Cal., No. 24-cv-5417, hearing 9/8/25.
This article does not necessarily reflect the opinion of Bloomberg Industry Group, Inc., the publisher of Bloomberg Law, Bloomberg Tax, and Bloomberg Government, or its owners.
Author Information
Nicholas J. Schneider is a member (partner) with Eckert Seamans’ intellectual property litigation and commercial litigation groups.
Write for Us: Author Guidelines
To contact the editors responsible for this story:
Learn more about Bloomberg Law or Log In to keep reading:
Learn About Bloomberg Law
AI-powered legal analytics, workflow tools and premium legal & business news.
Already a subscriber?
Log in to keep reading or access research tools.