Anthropic agrees to $1.5 billion settlement in largest copyright case
Anthropic settles copyright lawsuit for $1.5 billion after authors alleged piracy from Library Genesis and Pirate Library Mirror for AI training purposes.

Anthropic PBC has agreed to pay at least $1.5 billion to settle a copyright infringement lawsuit filed by authors who alleged the artificial intelligence company illegally used pirated copies of their books to train large-language models, according to court documents filed on September 5, 2025.
According to Justin Nelson, a lawyer representing the authors, "If approved, this landmark settlement will be the largest publicly reported copyright recovery in history, larger than any other copyright class action settlement or any individual copyright case litigated to final judgment."
The lawsuit, filed in federal court in California in August 2024, centered on roughly 500,000 published works. The proposed settlement amounts to a gross recovery of $3,000 per work, Nelson said in a memorandum to the judge in the case.
Subscribe PPC Land newsletter ✉️ for similar stories like this one. Receive the news every day in your inbox. Free of ads. 10 USD per year.
Piracy allegations drive massive settlement
The case emerged from allegations that Anthropic had "committed large-scale copyright infringement" by downloading and "commercially exploiting" books from pirating websites including Library Genesis and Pirate Library Mirror. According to court documents, the plaintiffs alleged Anthropic used these materials to train Claude, its artificial intelligence chatbot service.
Anthropic had argued its actions fell under "fair use" provisions of copyright law. In late June 2025, federal Judge William Alsup ruled that Anthropic's training methods constituted fair use because the end result was "exceedingly transformative." However, the ruling included crucial limitations.
Judge Alsup declared that downloading pirated copies of books did not constitute fair use. According to the court order, the judge found Anthropic's use of pirated material "inherently, irredeemably infringing."
"In June, the District Court issued a landmark ruling on AI development and copyright law, finding that Anthropic's approach to training AI models constitutes fair use," said Aparna Sridhar, deputy general counsel of Anthropic. "Today's settlement, if approved, will resolve the plaintiffs' remaining legacy claims."
Three-author lawsuit becomes industry precedent
The lawsuit was originally filed by three writers: Andrea Bartz, Charles Graeber, and Kirk Wallace Johnson. Bartz is a journalist and novelist, while Graeber and Johnson are journalists who have published nonfiction books. According to court records, these authors represented a potential class of up to 7 million claimants whose copyrighted works span over a century of publishing history.
The settlement structure includes four payments into the settlement fund, starting with a $300 million payout due within five business days of court approval, according to Nelson's filing. Anthropic will pay the full $1.5 billion within less than two years following final approval and will accelerate payments if it achieves certain milestones.
According to the settlement agreement, Anthropic must destroy books downloaded from Library Genesis or Pirate Library Mirror and any copies originating from the torrented versions. The company will also pay an additional $3,000 per work for any work Anthropic adds to the Works List above 500,000.
Buy ads on PPC Land. PPC Land has standard and native ad formats via major DSPs and ad platforms like Google Ads. Via an auction CPM, you can reach industry professionals.
AI industry faces mounting legal pressure
The settlement could significantly impact the trajectory of other pending litigation between AI platforms and published authors. PPC Land reported on the split decision in June 2025, where Judge Alsup found fair use for training but rejected piracy defenses.
John Grisham, "Game of Thrones" author George R.R. Martin, and Jodi Picoult are part of a group of nearly 20 bestselling authors who have sued OpenAI, alleging "systematic theft on a mass scale" for using their works to train ChatGPT and other tools.
Multiple AI companies face similar challenges. Reddit sued Anthropic in June 2025 over alleged unauthorized AI training on platform data. Ziff Davis filed a major lawsuit against OpenAI in April 2025, accusing the company of unauthorized use of content from 45 properties including CNET, IGN, and Mashable.
However, court decisions have not uniformly favored plaintiffs. Meta won a summary judgment motion in June 2025 when Judge Vince Chhabria ruled that using copyrighted books to train Llama large language models constituted fair use, though that ruling applied only to specific plaintiffs.
Technical details of AI training disputed
According to court documents, Anthropic assembled a central library containing both pirated and purchased books, then created various data mixes from these sources to train different large language model versions. The Northern District of California court examined each use separately under the four-factor fair use test.
The court found the training process "exceedingly transformative" and constituted fair use under Section 107 of the Copyright Act. "The purpose and character of using copyrighted works to train LLMs to generate new text was quintessentially transformative," the court stated.
However, according to court documents, Anthropic retained pirated copies even after determining they would not be used for training. The company planned to "store everything forever" with "no compelling reason to delete a book." The court distinguished between copies immediately transformed for fair use and those maintained in a general-purpose library.
According to the ruling, downloading from pirate sites when books could have been purchased lawfully was "inherently, irredeemably infringing." The piracy portion of the case was scheduled to proceed to trial in December 2025 to determine actual or statutory damages, including potential willfulness penalties.
Financial constraints drive settlement
According to court documents, Chief Financial Officer Krishna Rao stated that Anthropic appears "very unlikely" to obtain an appeal bond for the total possible amount of statutory damages should the company be found liable. This financial constraint creates what the company describes as coercive settlement pressure.
According to the Copyright Act, willful infringement can result in statutory damages up to $150,000 per work. With roughly 500,000 works in the class, potential damages could have reached billions of dollars if the case proceeded to trial.
The settlement compares favorably to other copyright class action recoveries. According to court documents, in Ferrick v. Spotify USA Inc., Judge Alison J. Nathan praised a class-wide copyright settlement valued at approximately $112.5 million as a "significant recovery" for the 535,000-member class, even though the award was valued at $210.28 per class member.
Broader copyright implications for AI development
The case reflects broader tensions within the AI industry regarding content licensing and fair use. Companies have adopted divergent strategies, with some pursuing formal licensing agreements while others rely on fair use doctrine to justify training on copyrighted materials.
The US Copyright Office released major guidance in May 2025 addressing when AI developers need permission to use copyrighted works. The report suggested that transformativeness and market effects would be the most significant factors in fair use determinations.
Google established partnerships with news organizations like The Associated Press, while OpenAI faces lawsuits from publishers including The New York Times. Concord Music and other publishers reached a partial agreement with Anthropic in January 2025 regarding copyright protection measures.
The Consumer Technology Association warned that allowing copyright class actions in AI training cases could result in a future where "copyright questions remain unresolved and the risk of 'emboldened' claimants forcing enormous settlements will chill investments in AI."
According to industry analysts, the settlement establishes a precedent that AI companies will face significant financial consequences for using pirated content in training datasets, even when the training itself might qualify as fair use.
Strategic business calculations behind the settlement
The settlement reflects sophisticated financial calculations by Anthropic about litigation risks and business implications. According to court documents, Chief Financial Officer Krishna Rao stated that Anthropic appears "very unlikely" to obtain an appeal bond for the total possible amount of statutory damages should the company be found liable. With roughly 500,000 works in the class and potential statutory damages up to $150,000 per work for willful infringement, Anthropic faced exposure exceeding $75 billion.
This financial constraint created what the company describes as coercive settlement pressure that could prevent proper adjudication of important fair use questions. Legal experts note that the $1.5 billion settlement represents approximately 2% of the maximum potential damages exposure, making it an economically rational decision despite the precedent it sets.
According to court filings, the settlement includes provisions for accelerated payments if Anthropic achieves certain business milestones. This structure suggests the company expects significant revenue growth that would make the settlement payments manageable relative to its business trajectory.
Market precedent and competitive implications
The settlement establishes a new cost structure for AI development using copyrighted materials. According to industry analysts, the $3,000 per work formula provides a baseline for licensing negotiations across the AI sector. This price point could influence how other companies approach content acquisition for training datasets.
PPC Land reported that industry groups warned allowing copyright class actions in AI training cases could result in "emboldened claimants forcing enormous settlements" that would chill AI investments. The Consumer Technology Association specifically argued that such potential liability "exerts incredibly coercive settlement pressure" that could stifle innovation.
According to Matthew Prince, CEO of Cloudflare, commenting on the settlement, the payment represents what it costs "to keep a judgment you like the dominant legal precedent." This observation highlights how companies may view settlements as strategic investments in shaping favorable legal frameworks rather than simple damage payments.
The settlement's timing coincides with increasing differentiation in AI company strategies. While Anthropic chose to settle, Meta won a summary judgment motion in June 2025 when Judge Vince Chhabria ruled that using copyrighted books to train large language models constituted fair use, though that ruling applied only to specific plaintiffs.
Economic impact on content creation industry
The settlement addresses fundamental questions about economic relationships between AI companies and content creators. According to court documents, plaintiffs alleged that "Anthropic, in taking authors' works without compensation, has deprived authors of books sales and licensing revenues." They further argued that "there has long been an established market for the sale of books and e-books," yet Anthropic "ignored it and chose to scrape a massive corpus of copyrighted books from the internet."
Authors Guild CEO Mary Rasenberger praised the settlement as "a strong message to the AI industry that there are serious consequences when they pirate authors' works to train their AI, robbing those least able to afford it." According to her statement, the settlement provides compensation to "authors, publishers, and rightsholders generally."
AAP CEO and former Register of Copyrights Maria Pallante called the settlement "beneficial to all class members," noting that it "provides enormous value in sending the message that Artificial Intelligence companies cannot unlawfully acquire content from shadow libraries or other pirate sources to use as the building blocks for their businesses."
The settlement structure includes detailed notice provisions targeting authors and publishers across multiple platforms. According to court documents, notice will be distributed through Google Display Network, Facebook, Instagram, and Reddit, specifically targeting adults interested in book publishing, copyright law, and AI training data topics.
Legal precedent and future litigation strategy
According to legal experts, the settlement creates important distinctions for future AI copyright cases. The agreement releases only past claims and explicitly excludes claims based on AI model outputs. According to court documents, "no claims based on the output of AI models are released" and "there will be no release of any claims for future reproduction, distribution and/or creation of derivative works."
This limited scope suggests that while the settlement resolves historical training practices, it does not provide comprehensive protection for AI companies against future copyright claims. The agreement specifically states that "the release does not constitute, in any respect, a license to torrent, scan, or train AI models on any copyrighted works, or to create infringing outputs from AI models."
The settlement distinguishes between different sources of training data. According to court documents, Judge Alsup found that reproducing purchased books to train AI constituted fair use, while downloading from pirate sites when books could have been purchased lawfully was "inherently, irredeemably infringing." This legal framework suggests future litigation will focus on data acquisition methods rather than the training process itself.
According to industry analysts, the settlement's focus on pirated sources rather than training methodology could encourage AI companies to invest more heavily in legitimate content licensing while potentially providing cover for fair use arguments regarding properly acquired materials.
Subscribe PPC Land newsletter ✉️ for similar stories like this one. Receive the news every day in your inbox. Free of ads. 10 USD per year.
Timeline
- August 2024: Authors Andrea Bartz, Charles Graeber, and Kirk Wallace Johnson file putative class action lawsuit
- October 2024: Court requires class certification motions by March 6, 2025, setting accelerated litigation schedule
- January 2025: Concord Music and other publishers reach partial agreement with Anthropic on copyright protection measures
- February 2025: Court grants Anthropic's motion for early summary judgment on fair use before class certification
- April 2025: Anthropic claws back spreadsheet showing data mix compositions used for training various LLMs
- May 2025: US Copyright Office releases comprehensive AI training guidance addressing fair use questions
- June 2025: Reddit sues Anthropic over unauthorized AI training on platform data
- June 23, 2025: Judge William Alsup issues mixed ruling on fair use and piracy claims in landmark split decision
- July 31, 2025: Anthropic appeals largest copyright class action certification seeking permission to appeal
- September 5, 2025: Settlement agreement filed with court proposing $1.5 billion payment
Subscribe PPC Land newsletter ✉️ for similar stories like this one. Receive the news every day in your inbox. Free of ads. 10 USD per year.
Summary
Who: Anthropic PBC, an artificial intelligence company, and authors Andrea Bartz, Charles Graeber, and Kirk Wallace Johnson representing a class of up to 7 million potential claimants whose copyrighted books were allegedly used without authorization.
What: A $1.5 billion settlement agreement resolving copyright infringement claims over Anthropic's use of pirated books from Library Genesis and Pirate Library Mirror to train its Claude AI models. The settlement amounts to approximately $3,000 per work for roughly 500,000 copyrighted books.
When: The settlement was filed on September 5, 2025, following a lawsuit filed in August 2024 and a split court ruling on June 23, 2025, that found fair use for AI training but rejected piracy defenses.
Where: The case was decided in the US District Court for the Northern District of California, with Senior US District Judge William Alsup presiding over the landmark copyright and artificial intelligence case.
Why: The settlement resolves claims that Anthropic violated copyright law by downloading pirated books rather than purchasing legitimate copies, even though the court found the AI training process itself constituted transformative fair use under federal copyright law.