Anthropic finds 250 poisoned documents are enough to backdoor large language models

2025-10-10 05:35 GMT · 6 months ago aimagpro.com

Anthropic, working with the UK’s AI Security Institute and the Alan Turing Institute, has discovered that as few as 250 poisoned documents are enough to insert a backdoor into large language models – regardless of model size.
The article Anthropic finds 250 poisoned documents are enough to backdoor large language models appeared first on THE DECODER.