#scaling-effects

[ follow ]
Artificial intelligence
fromInfoQ
2 days ago

Anthropic Finds LLMs Can Be Poisoned Using Small Number of Documents

Injecting about 250 poisoned pretraining documents can implant a backdoor causing gibberish outputs, and such poisoning becomes easier as models scale.
[ Load more ]