Meta wants the courts to declare that piracy is legal—as long as the pirates are training AI.

In a legal filing that’s sent shockwaves through the creative community, Meta (Facebook’s parent company) has argued that downloading copyrighted books via BitTorrent qualifies as fair use under U.S. copyright law. Their reasoning? The books were used to train AI language models, which constitutes “transformative” use.

If Meta wins this argument, it could reshape not just AI development, but the entire foundation of digital copyright—and your rights as a creator.

What Meta Is Actually Arguing

In ongoing copyright litigation, Meta faces claims that it illegally used copyrighted books to train its AI models. Rather than denying they used the content, Meta’s legal team is taking a bold stance:

Yes, we used pirated books. Yes, we obtained them via BitTorrent. And yes, that’s perfectly legal.

Their argument hinges on the four-factor fair use test:

  1. Purpose and character of use: Meta claims AI training is “transformative”—the books aren’t reproduced as books, but digested into statistical patterns
  2. Nature of the copyrighted work: Creative works normally get more protection, but Meta argues the use is so different it doesn’t matter
  3. Amount used: They used entire books, but claim this was necessary for AI training
  4. Market effect: Meta argues AI models don’t compete with the books themselves

This is an audacious legal strategy. They’re not disputing the piracy—they’re arguing the piracy doesn’t matter.

The “Transformative Use” Stretch

Fair use law allows certain uses of copyrighted material without permission. The most protected uses are “transformative”—they don’t just copy the work, they create something new.

Classic examples:

  • Parody (a comedian mocking a song)
  • Commentary (a reviewer quoting a book)
  • Scholarship (an academic analyzing a text)

Meta is stretching this concept beyond recognition. Their argument essentially says:

“We didn’t copy the books to read them. We copied them to feed them into a statistical model that learned patterns. The model doesn’t contain the books—it contains patterns derived from the books. Therefore, our use is transformative.”

If this logic holds, any computational processing of copyrighted material becomes fair use. Download a movie to analyze its color palette? Fair use. Pirate music to train a recommendation algorithm? Fair use. Scrape an entire news website to build a summarization tool? Fair use.

Why BitTorrent Makes This Worse

Meta didn’t just use copyrighted books—they obtained them via BitTorrent, a peer-to-peer file sharing network primarily used for piracy. This adds layers of legal and ethical problems:

BitTorrent Requires Re-Distribution

When you download via BitTorrent, you simultaneously upload pieces to others. Meta wasn’t just downloading—they were actively distributing pirated content to other users.

The Source Was Obviously Illegal

BitTorrent book collections like LibGen and Z-Library are notorious piracy repositories. Meta can’t claim innocent mistake—they went to the digital equivalent of a bootleg DVD market and loaded up trucks.

It Demonstrates Intent

Going to BitTorrent for training data, rather than licensing content legally, shows a deliberate choice to use pirated material when legal alternatives existed.

What This Means for Authors and Creators

If Meta’s argument succeeds, the implications are grim:

Your Work Can Be Pirated for AI Training

Any book, article, image, song, or video can be copied without permission as long as it’s being fed into an AI model. The “transformative” magic wand makes piracy legal.

Unlike licensing agreements, fair use doesn’t require payment or permission. Authors whose books train AI models get nothing—not a cent, not a credit, not even notification.

Impossible to Opt Out

How do you prevent your work from being used if piracy is legal? You can’t. Once your book exists in digital form, it can be scraped, copied, and processed without recourse.

The Value of Creation Diminishes

Why pay for content when you can train AI on pirated copies? The economic incentive for creating original work erodes when AI companies can legally use everything for free.

The Privacy Angle: Your Data, Their Model

This case has profound privacy implications beyond copyright:

Data Provenance Is Impossible

If AI companies can use pirated content, tracking what’s in a model becomes impossible. Was your personal blog post in the training data? Your emails? Your medical records? You’ll never know.

Privacy laws increasingly emphasize consent—you should control how your data is used. But if AI training is fair use, your consent is irrelevant. They can use it anyway.

AI Models as Derivative of You

When an AI model learns from your writing, it captures something of your style, your knowledge, your voice. That model is, in a sense, derived from you. Meta’s argument says they own that derivative without owing you anything.

What Other AI Companies Are Doing

Meta isn’t alone in using questionable training data, but approaches vary:

OpenAI has faced similar lawsuits (including from the New York Times) and has generally settled or sought licensing deals rather than making the aggressive fair use argument.

Anthropic (maker of Claude) has been more cautious about training data sources and has emphasized constitutional AI approaches that try to minimize harm.

Google has extensive licensing agreements for some content categories but has also faced criticism for data scraping practices.

Stability AI has been sued over image generation models trained on copyrighted artwork, with similar fair use defenses.

Meta’s argument is the most aggressive—essentially asking courts to legalize piracy for AI training across the board.

What Happens If Meta Wins

A Meta victory would:

  1. Legitimize AI data laundering: Pirate content, process it through AI, claim the output is “transformative”
  2. Kill licensing markets: Why pay for training data when piracy is free and legal?
  3. Accelerate the AI copyright crisis: Every creator becomes an unwilling, uncompensated AI trainer
  4. Set international precedent: Other countries may follow U.S. fair use interpretations
  5. Undermine trust in AI: Knowing models are built on piracy changes how we view them

What You Can Do

Support Author Organizations

Groups like the Authors Guild are fighting these cases. They need resources and visibility.

Contact Your Representatives

Copyright law is ultimately made by Congress. Let them know creators’ rights matter.

Be Selective About Platforms

Some AI companies are more ethical about training data than others. Your choices matter.

Understand Your Rights

Even if fair use is abused, other rights may apply. Privacy laws, terms of service, and contract law offer alternative protections.

Document Your Work

Proving your content was used in training data is hard. Maintain records of your publications for potential future claims.

The Bottom Line

Meta’s argument—that BitTorrent piracy is fair use for AI training—represents either a bold legal innovation or a brazen attempt to legalize theft at scale. The outcome will shape the future of AI development, creator rights, and digital privacy for decades.

The courts will decide whether “we pirated it for AI” is a valid defense. But regardless of the legal outcome, the ethical reality is clear: taking creators’ work without permission or compensation isn’t innovation. It’s exploitation.

And dressing it up in transformative use language doesn’t change what it is.


Protecting your digital rights starts with understanding them. Follow My Privacy Blog for news and analysis on privacy, data rights, and AI ethics.