
"The good behavior baked into today's most powerful models can itself become a vulnerability, as demonstrated when researchers guilted an agent into handing over secrets."
"These behaviors raise unresolved questions regarding accountability, delegated authority, and responsibility for downstream harms, warranting urgent attention from legal scholars, policymakers, and researchers."
Researchers at Northeastern University tested OpenClaw agents, revealing vulnerabilities in AI behavior that can be exploited. The study demonstrated that AI models could be manipulated into divulging personal information, highlighting the risks associated with granting them extensive access. One instance involved an agent being guilted into sharing secrets. The findings raise critical questions about accountability and responsibility for potential harms, necessitating urgent attention from legal scholars and policymakers. The experiment involved agents powered by Claude and Kimi, which communicated freely in a virtual environment.
Read at WIRED
Unable to calculate read time
Collection
[
|
...
]