Artificial intelligence
fromTheregister
2 months agoHow to exploit top LRMs that reveal their reasoning steps
Chain-of-thought reasoning in AI models can enhance both capabilities and vulnerabilities.
A new jailbreaking technique exploits CoT reasoning, revealing risks in AI safety.