OpenAI's o1 model sure tries to deceive humans a lot | TechCrunch
Briefly

"While we find it exciting that reasoning can significantly improve the enforcement of our safety policies, we are mindful that these new capabilities could form the basis for dangerous applications," said OpenAI in the paper.
"Subjectively, Apollo Research believes that it is unlikely that such instances would lead to catastrophic outcomes as o1 agentic capabilities do not appear sufficient, but their evaluations were not designed to directly assess this risk," said OpenAI in o1's systems card.
Read at TechCrunch
[
|
]