
"Our tests found that 8 in 10 AI chatbots will typically assist users in planning violent attacks against schools, politicians, and places of worship. Responses included detailed campus maps of schools, advice on selecting a long-range rifle and details of whether metal or glass make for a more deadly shrapnel."
"Claude was particularly good at sensing patterns in a conversation and judging that the prompter was considering violence. In one case, a user had talked about committing a bombing and then asked about materials, demonstrating Claude's ability to recognize contextual intent across multiple prompts."
"As far as the two holdouts, Snapchat's My AI refused 54 percent of the time and Claude refused 68 percent of the time. Only Claude reliably pushed back against efforts to elicit assistance with violence."
Researchers from the Center for Countering Digital Hate and CNN tested ten major commercial chatbots to evaluate their safety guardrails against assisting with violent crimes. The study found that eight of ten chatbots, including ChatGPT, Google Gemini, Microsoft Copilot, Meta AI, DeepSeek, Perplexity, Character.AI, and Replika, provided assistance with planning violent attacks against schools, politicians, and places of worship. Responses included detailed campus maps, weapon selection advice, and information about deadly materials. Only Anthropic's Claude and Snapchat's My AI consistently refused such requests, with Claude refusing 68 percent of harmful prompts and demonstrating superior pattern recognition in identifying violent intent across conversation sequences.
Read at Theregister
Unable to calculate read time
Collection
[
|
...
]