Keep Factually independent
Whether you agree or disagree with our analysis, these conversations matter for democracy. We don't take money from political groups - even a $5 donation helps us keep it that way.
Fact check: Justice william alsup
Executive Summary
Judge William Alsup has ruled that using purchased copyrighted books to train large language models can be fair use, while also finding that assembling or using pirated copies for a centralized digital library may not be protected—leaving mixed results and open questions about model outputs and downstream uses [1]. The decisions, issued in mid-2025, favor AI developers on training-copy issues but preserve liability avenues where unauthorized copying or distribution occurred, and they leave core market-harm questions unresolved [1] [2].
1. A Clear Claim: Training Equals Fair Use—But with Limits That Matter
Judge Alsup’s rulings adopt a transformative-use frame, finding that copying millions of books to train Anthropic’s Claude LLM was “exceedingly transformative” and thus favored fair use, because the process created a machine-learning model rather than a substitute product for the books themselves [1]. Alsup explicitly distinguished the act of ingestion for model training from the creation of a retrievable, centralized digital library: the latter resembled distribution and preservation of text copies and was therefore not automatically protected by the same fair-use rationale, creating a legal split within the same case [1] [3].
2. The Ruling’s Legal Logic and Evidence Emphasis
Alsup’s opinion centers on transformativeness and the absence of demonstrated market harm to publishers for training copying; he emphasized that training creates a different product category—an AI model—that does not function as a market substitute for books, and that plaintiffs had not supplied convincing empirical proof of market displacement [2]. The judge’s approach ties the outcome to evidentiary showings; in other words, the decision turns on both doctrinal framing and plaintiffs’ failure to marshal strong empirical evidence linking training copies to lost sales or markets [2].
3. What Alsup Did Not Resolve: Outputs and Downstream Uses
Alsup’s rulings leave open a critical question: whether specific outputs of generative AI—verbatim passages, close paraphrases, or stylized imitations—constitute copyright infringement or fair use. The decision resolves the lawfulness of ingestion but not the downstream legality of generated content, which he flagged as a separate inquiry likely to depend on detailed factual records at trial or in future appeals [2]. This gap means companies can cite the decision to support training activities, but risk remains around product responses and commercial deployment.
4. Piracy Versus Lawful Acquisition: A Sharp Line Drawn
Alsup drew a bright-line distinction between training on lawfully acquired copies and using pirated or unauthorized repositories to create a searchable library: the latter exposed Anthropic to potential liability because it resembled classic distribution or preservation of unauthorized copies rather than opaque model ingestion [1] [3]. The court declined to bless the use of pirated corpora wholesale and left unresolved factual claims about how some copies were obtained, signaling that lawfulness of acquisition remains a central determinant of liability [1].
5. Industry and Advocacy Reactions: Winners, Warnings, and Strategic Narratives
AI companies and industry groups treated Alsup’s rulings as a major legal win for training practices, citing the decision as precedent that ingestion can be fair use; conversely, authors, publishers, and copyright advocates highlighted Alsup’s refusal to greenlight pirated libraries and stressed unresolved risks around outputs and market effects [1] [2]. Both sides have incentives: the tech sector seeks broad training immunity, while content owners emphasize downstream harms and procedural deficiencies in evidence that courts must fill in future actions.
6. Procedural Posture: Trials, Appeals, and the Road Ahead
Alsup’s mixed rulings set up predictable next steps: factual trials on the provenance of disputed copies and market-harm evidence, and likely appeals that could reach higher courts because of the high stakes for both the AI industry and content owners [3] [2]. The rulings create a patchwork posture—training may enjoy provisional judicial endorsement, but concrete liability can attach if factual records show piracy or demonstrable market substitution—so litigation and regulatory attention will continue to shape precedent.
7. Broader Legal and Policy Stakes: Evidence, Markets, and Judicial Method
The decisions underscore the judiciary’s reliance on empirical proof to distinguish transformative training from infringing distribution; judges will increasingly ask for market studies, usage logs, and provenance records. Alsup’s method emphasizes doctrinal nuance rather than sweeping rules, so policymakers and litigants will need to grapple with how to standardize discovery, protect trade secrets, and reconcile competing incentives between innovation and authorial remuneration [2]. The outcome will influence contract practices, licensing strategies, and potential legislative responses.
8. Bottom Line for Stakeholders: Cautious Opportunity and Continuing Risk
For AI developers, Alsup’s rulings provide a useful precedent supporting training on lawfully obtained copyrighted works—but they do not offer blanket immunity, especially where acquisition practices are suspect or model outputs replicate works. For creators and publishers, the decisions preserve pathways to challenge downstream copying and distribution claims, provided they can develop evidence of market impact or unauthorized copying. The net effect is conditional legal advantage, not definitive clearance, ensuring litigation and regulation will continue to define the field [1].