Artificial intelligence
fromInfoQ
2 days agoAnthropic Finds LLMs Can Be Poisoned Using Small Number of Documents
Injecting about 250 poisoned pretraining documents can implant a backdoor causing gibberish outputs, and such poisoning becomes easier as models scale.