
"The model, DeepSeekMath-V2, scored 118 out of 120 points on questions from the 2024 William Lowell Putnam Mathematical Competition, beating the top human score of 90. The model also performed at the level of gold-medal winners in the International Mathematical Olympiad (IMO) 2025 and the 2024 China Mathematical Olympiad. The results are described in a preprint posted on arXiv on 27 November."
"Chinese artificial intelligence company DeepSeek has released a mathematical reasoning model that can identify and correct its own errors. The model beat the best human score in one of the world's most prestigious undergraduate maths competitions. "We are at a point where AI is about as good at maths as a smart undergraduate student," says Kevin Buzzard, a mathematician at Imperial College London. "It is very exciting.""
DeepSeekMath-V2 achieved 118 out of 120 on the 2024 William Lowell Putnam Mathematical Competition, surpassing the top human score of 90. The model also performed at gold-medal levels in the International Mathematical Olympiad 2025 and the 2024 China Mathematical Olympiad. The model introduces self-verifiable mathematical reasoning, incorporating a verifier trained to evaluate mathematical proofs and identify and correct its own errors. Prior AI solvers such as AlphaGeometry 2 and Gemini's Deep Think also reached gold-level IMO performance earlier in the year. The approach emphasizes rewarding sound intermediate reasoning rather than only final-answer accuracy.
Read at Nature
Unable to calculate read time
Collection
[
|
...
]