#adversarial-attacks

[ follow ]
fromFortune
3 days ago

AI's ability to 'think' makes it more vulnerable to new jailbreak attacks, new research suggests | Fortune

Using a method called "Chain-of-Thought Hijacking," the researchers found that even major commercial AI models can be fooled with an alarmingly high success rate, more than 80% in some tests. The new mode of attack essentially exploits the model's reasoning steps, or chain-of-thought, to hide harmful commands, effectively tricking the AI into ignoring its built-in safeguards. These attacks can allow the AI model to skip over its safety guardrails and potentially
Artificial intelligence
fromIT Pro
4 days ago

Some of the most popular open weight AI models show 'profound susceptibility' to jailbreak techniques

A host of leading open weight AI models contain serious security vulnerabilities, according to researchers at Cisco. In a new, researchers found these models, which are publicly available and can be downloaded and modified by users based on individual needs, displayed "profound susceptibility to adversarial manipulation" techniques. Cisco evaluated models by a range of firms including: Alibaba (Qwen3-32B) DeepSeek (v3.1) Google (Gemma 3-1B-IT) Meta (Llama 3.3-70B-Instruct) Microsoft (Phi-4) OpenAI (GPT-OSS-20b) Mistral (Large-2).
Artificial intelligence
#lidar
fromFuturism
1 month ago
Miscellaneous

Putting Mirrors on Traffic Cones Causes Self-Driving Cars to Melt Down, Confounding Lidar

fromFuturism
1 month ago
Miscellaneous

Putting Mirrors on Traffic Cones Causes Self-Driving Cars to Melt Down, Confounding Lidar

Artificial intelligence
fromNextgov.com
2 months ago

Is artificial intelligence a friend, foe or frenemy? NIST wants to find out

AI's dual-use nature accelerates both cybersecurity defenses and adversary attacks, creating urgent need for resilience and guidance from standards bodies like NIST.
[ Load more ]