#jailbreaking-techniques

[ follow ]
Artificial intelligence
fromTheregister
2 months ago

How to exploit top LRMs that reveal their reasoning steps

Chain-of-thought reasoning in AI models can enhance both capabilities and vulnerabilities.
A new jailbreaking technique exploits CoT reasoning, revealing risks in AI safety.
Loading...