Anthropic Research: About 250 Poisoned Documents Can Backdoor an LLM, Model Size Does Not Matter
Anthropic research shows approximately 250 malicious documents are sufficient to implant a backdoor in an LLM, and the required number is independent of model parameters (consistent from 600M to 13B). This challenges the assumption that larger models are harder to poison.