ARTICLE
16 October 2025

Small Data Poisoning Can Compromise Large AI Models

Anyone deploying AI in high-stakes environments should pay attention to a recent paper from Anthropic, the Alan Turing Institute, et al.
United States Technology
Michael Best are most popular:
  • within Immigration, Corporate/Commercial Law and Finance and Banking topic(s)

Anyone deploying AI in high-stakes environments should pay attention to a recent paper from Anthropic, the Alan Turing Institute, et al. Anthropic's study shows that injecting just 250 poisoned samples—out of potentially billions of training tokens—can reliably compromise even the largest language models. This means that less than a fraction of a fraction of the data can embed persistent vulnerabilities, making poisoning not only feasible but potentially dangerously efficient. More troubling, the study indicates that increasing model size does not mitigate this risk. The attack cost remains relatively low regardless of scale.

For law firms leveraging AI to support research, drafting, or client services, this finding underscores the importance of understanding where vulnerabilities may lie. Many AI applications are little more than specialized front ends to general-purpose LLMs. The applications use engineered prompts, context management protocols, and workflow orchestration to elicit customized outputs from these generic models. However, once the core model is compromised, protective layers around it have very limited corrective power, and applications using the model remain at risk until the foundational issue is rectified.

What can law firms and other professional AI users do in response to these findings?

One promising answer is to consider smaller, custom-trained models built for specific tasks. These models can be built on well-audited, curated datasets and aligned to specific legal workflows, minimizing exposure to poisoning attacks. They're also better aligned to the workflows they're designed to support, reducing the need for complex prompt engineering and minimizing unpredictable behavior.

From a governance perspective, small models offer clearer provenance, tighter control, and more defensible compliance postures. They're easier to align with privacy regulations like GDPR and HIPAA, and they reduce the risk of embedding copyrighted or sensitive material.

As AI adoption accelerates, the temptation to use massive general-purpose models for everything is strong. But Anthropic's findings suggest that bigger isn't always better. For sensitive applications, precision and trust matter more than scale. Tailoring AI solutions to specific needs by building small, purpose-driven models may be the smarter—and safer—path forward.

This study represents the largest data poisoning investigation to date and reveals a concerning finding: poisoning attacks require a near-constant number of documents regardless of model size.

www.anthropic.com/...

The content of this article is intended to provide a general guide to the subject matter. Specialist advice should be sought about your specific circumstances.

Mondaq uses cookies on this website. By using our website you agree to our use of cookies as set out in our Privacy Policy.

Learn More