For the last several years, the intellectual property community has placed a heavy focus on the “input” side of generative artificial intelligence (AI). Widespread media coverage has highlighted massive lawsuits questioning whether technology companies may legally scrape books, artwork, and music to train their commercial models.
As court rulings emerge, a major shift is occurring. Federal judges are establishing clear distinctions between the source of the training data and the specific outputs the AI generates.
Whether operating a business that utilizes generative AI to optimize workflows or representing a creator looking to protect original work, it’s vital to understand this change. The legal battlefield is moving past the mechanics of how an AI learned and is instead scrutinizing data sourcing and actual AI outputs.
The Core Distinction: Lawful Data vs. Pirated Sources
In several major rulings emerging from California federal courts, judges have now established preliminary ground rules regarding how the "fair use" doctrine applies to AI training data.
For the most part, courts have been open to the argument that training models on lawfully acquired data constitute fair use. Judges recognize that these models do not only store copies of art and text but rather learn statistical patterns to develop entirely new content.
However, courts are now drawing a hard line regarding unlawfully acquired content. If a model was trained on pirated books or compromised databases, judges warn that it constitutes a severe compliance failure. Consequently, companies developing or fine-tuning their own models face significant legal and compliance risks if they are not diligent about the provenance of their training data.
As summarized in an overview of copyright law by Baker Donelson, this distinction highlights a substantial compliance risk for companies developing or fine-tuning their own models.
The High Burden of Proof for Infringing Outputs
While the litigation surrounding data sourcing continues, courts are demanding much more rigorous proof regarding the "outputs" generated by AI platforms.
Early on, creators argued that an AI product was automatically an illegal derivative work simply because it was trained on their copyrighted material. Federal judges have largely rejected this broad theory.
Instead, a growing consensus among federal judges requires plaintiffs to prove that a specific AI output is substantially similar to their copyrighted work. For creators, it is no longer sufficient to demonstrate that a book or painting was included in the training set. To move forward with an output-driven lawsuit, plaintiffs must show that the AI actually generated an expressive work that mirrors their protected material.
Track the active case status of Andersen v. Stability AI, where visual artists have survived motions to dismiss and are actively proceeding through the discovery phase.
Proving Measurable Economic Harm
Under traditional copyright law, the "fair use" analysis leans heavily on whether the secondary work damages the potential market for the original. This continues to be a fiercely debated topic in active AI litigation.
Several judges have noted that because AI can flood the market with synthetic content at a massive scale, it poses a unique threat to creators. However, courts are making it clear that creators cannot rely on speculative harm. To prevail, plaintiffs must show concrete evidence that AI outputs are directly competing with or replacing the market for their original work.
Actionable Steps for Businesses and Creators
With the legal landscape heavily focused on proper data sourcing and identical outputs, proactive risk management is necessary.
For Businesses Using Generative AI:
- Verify Your Data Sources: If you are training or fine-tuning AI models, ensure all training data is legally acquired and fully licensed.
- Audit Prompts and Workflows: Instruct your teams to avoid using prompts that explicitly request an AI to copy the style of a specific living artist or replicate existing copyrighted works.
- Implement Output Filtering: Ensure the AI tools your enterprise relies on have guardrails in place that block the generation of near-identical text or heavily sampled imagery.
- Review Vendor Contracts: Read your AI service agreements carefully. Ensure your vendor provides clear intellectual property indemnification that covers both the training data and the outputs the tool creates.
For Content Creators and IP Holders:
- Monitor for Infringement: Utilize digital watermarking and active monitoring tools to track down synthetic content that directly copies or heavily borrows from your work.
- Focus on the Output: If you decide to take legal action against an AI platform, your strategy should be focused on gathering solid proof of identical outputs and direct market displacement.
Navigating the Future of IP in the AI Era
The intersection of AI and copyright law is actively being defined in the courts. As judges crack down on pirated training data and demand proof of real similarity for outputs, navigating this landscape requires a proactive strategy.
Whether you need to defend your branding from generative look-alikes or ensure your company's use of AI is legally compliant, Bochner PLLC is here to offer strategic legal counsel.