OpenAI and Anthropic evaluated each others' models - which ones came out on top
Briefly

OpenAI and Anthropic ran internal safety and misalignment evaluations on each other's models, focusing on alignment, sycophany, and hallucinations to reveal performance gaps. Both labs published detailed reports outlining their findings and identifying blind spots that the other company had missed. The swap demonstrates competitors collaborating to identify risks and strengthen model safety while balancing rapid product development with evaluation thoroughness. Analysts framed the collaboration as strategically necessary because shared risks from increasingly powerful AI outweigh benefits of unchecked competition, while also noting potential policy signaling by industry leaders. The teams compiled top insights and analysis from their cross-evaluations.
The AI race is in full swing, and companies are sprinting to release the most cutting-edge products. Naturally, this has raised concerns about speed compromising proper safety evaluations. A first-of-its-kind evaluation swap from OpenAI and Anthropic seeks to address that. Also: OpenAI used to test its AI models for months - now it's days. Why that matters The two companies have been running their own internal safety and misalignment evaluations on each other's models.
"This rare collaboration is now a strategic necessity. The report signals that for the AI titans, the shared risk of an increasingly powerful AI product portfolio now outweighs the immediate rewards of unchecked competition," said Gartner analyst Chirag Dekate. That said, Dekate also noted the policy implications, calling the reports "a sophisticated attempt to frame the safety debate on the industry's own terms, effectively saying, 'We understand the profound flaws better than you do, so let us lead.'"
Read at ZDNET
[
|
]