Legal risk, data rights, and competition dynamics are increasingly critical to measuring an AI system’s integrity. These legal shifts raise demand for third-party, rights-respecting evaluators.
Recent copyright settlements and antitrust fines mark potential turning points for how AI systems may be evaluated, governed, and trusted. From the perspective of AI evaluations, they signal that such evaluations can no longer focus exclusively on technical performance, or assessments of factuality, bias, and robustness, but must also account for data provenance, legal compliance, and market power.
Between September and October 2025, Anthropic agreed to pay approximately $1.5 billion to settle claims from authors and publishers that it had trained its large language model on pirate websites LibGen and Pirate Library Mirror (PiLiMi). A federal court had earlier held that training on legally purchased books might qualify as fair use, but using pirated libraries did not. As part of the settlement, Anthropic has agreed “to destroy all copies of pirated books in its possession” and pay rightsholders $3,000 per title. This case illustrates that the legality of AI training depends not just on the model’s purpose, but also on the lawful acquisition of that data. Other lawsuits against AI companies are also underway, including against OpenAI, similarly alleging copyright infringement.
During this time, the European Commission (EC) fined Google €2.95 billion (US $3.45 billion) for abusing dominance in its ad-tech business. The EC ordered the company to stop favoring its own ad exchange and warning that stronger remedies, including structural separation, could follow. While the Google case was not explicitly concerning AI, it reinforces that regulators, particularly those in Europe, are prepared to levy multi-billion-dollar penalties for anti-competitive or opaque practices in digital markets.
Together, these actions reshape the context for AI evaluation in four key ways:
For evaluators and policymakers, the message suggests that legal risk, data rights, and competition dynamics are increasingly critical to measuring an AI system’s integrity. These legal shifts raise demand for third-party, rights-respecting evaluators.
At the same time, the debate on copyright infringement remains a turbulent one, with competing developments. A U.K. court recently ruled in favor of Stability AI in a copyright case brought by Getty Images, offering competing precedent in the debate over AI training data. Getty had accused Stability of copying millions of its photos to train Stable Diffusion and of reproducing its watermark in generated images. The court dismissed all copyright claims, finding no evidence of direct copying, though it acknowledged some limited trademark issues. The decision is considered a boost for AI developers, suggesting that models trained on copyrighted material may avoid infringement if they do not reproduce works verbatim.
Ultimately, in the years ahead, trust in AI may depend as much on how models are constructed and governed, as on what they can do and output. Evaluation frameworks may evolve to include governance and compliance metrics, verifying lawful data sources, auditing licensing statements, and tracking vendor accountability. At the same time, the discussion remains a volatile one, and future legal decisions may shift the trajectory of accountability and correspondingly, AI evaluations.