Google’s AI Can Beat the Smartest High Schoolers in Math

Google’s AlphaGeometry2 AI reaches the level of gold-medal students in the International Mathematical Olympiad

Blue cube pyramid with blue sky background.

Google DeepMind’s AI AlphaGeometry2 aced problems set at the International Mathematical Olympiad.

Wirestock, Inc./Alamy Stock Photo

Join Our Community of Science Lovers!

A year ago AlphaGeometry, an artificial-intelligence (AI) problem solver created by Google DeepMind, surprised the world by performing at the level of silver medallists in the International Mathematical Olympiad (IMO), a prestigious competition that sets tough maths problems for gifted high-school students.

The DeepMind team now says the performance of its upgraded system, AlphaGeometry2, has surpassed the level of the average gold medallist. The results are described in a preprint on the arXiv.

“I imagine it won’t be long before computers are getting full marks on the IMO,” says Kevin Buzzard, a mathematician at Imperial College London.


On supporting science journalism

If you're enjoying this article, consider supporting our award-winning journalism by subscribing. By purchasing a subscription you are helping to ensure the future of impactful stories about the discoveries and ideas shaping our world today.


Solving problems in Euclidean geometry is one of the four topics covered in IMO problems — the others cover the branches of number theory, algebra and combinatorics. Geometry demands specific skills of an AI, because competitors must provide a rigorous proof for a statement about geometric objects on the plane. In July, AlphaGeometry2 made its public debut alongside a newly unveiled system, AlphaProof, which DeepMind developed for solving the non-geometry questions in the IMO problem sets.

Mathematical language

AlphaGeometry is a combination of components that include a specialized language model and a ‘neuro-symbolic’ system — one that does not train by learning from data like a neural network but has abstract reasoning coded in by humans. The team trained the language model to speak a formal mathematical language, which makes it possible to automatically check its output for logical rigour — and to weed out the ‘hallucinations’, the incoherent or false statements that AI chatbots are prone to making.

For AlphaGeometry2, the team made several improvements, including the integration of Google’s state-of-the-art large language model, Gemini. The team also introduced the ability to reason by moving geometric objects around the plane — such as moving a point along a line to change the height of a triangle — and solving linear equations.

The system was able to solve 84% of all geometry problems given in IMOs in the past 25 years, compared with 54% for the first AlphaGeometry. (Teams in India and China used different approaches last year to achieve gold-medal-level performance in geometry, but on a smaller subset of IMO geometry problems.)

The authors of the DeepMind paper write that future improvements of AlphaGeometry will include dealing with maths problems that involve inequalities and non-linear equations, which will be required to to “fully solve geometry.”

Rapid progress

The first AI system to achieve a gold-medal score for the overall test could win a US$5-million award called the AI Mathematical Olympiad Prize — although that competition requires systems to be open-source, which is not the case for DeepMind.

Buzzard says he is not surprised by the rapid progress made both by DeepMind and by the Indian and Chinese teams. But, he adds, although the problems are hard, the subject is still conceptually simple, and there are many more challenges to overcome before AI is able to solve problems at the level of research mathematics.

AI researchers will be eagerly awaiting the next iteration of the IMO in Sunshine Coast, Australia, in July. Once its problems are made public for human participants to solve, AI-based systems get to solve them, too. (AI agents are not allowed to take part in the competition, and are therefore not eligible to win medals.) Fresh problems are seen as the most reliable test for machine-learning-based systems, because there is no risk that the problems or their solution existed online and may have ‘leaked’ into training data sets, skewing the results.

This article is reproduced with permission and was first published on February 7, 2025.

Davide Castelvecchi is a staff reporter at Nature who has been obsessed with quantum spin for essentially his entire life. Follow him on X @dcastelvecchi

More by Davide Castelvecchi

First published in 1869, Nature is the world's leading multidisciplinary science journal. Nature publishes the finest peer-reviewed research that drives ground-breaking discovery, and is read by thought-leaders and decision-makers around the world.

More by Nature magazine

It’s Time to Stand Up for Science

If you enjoyed this article, I’d like to ask for your support. Scientific American has served as an advocate for science and industry for 180 years, and right now may be the most critical moment in that two-century history.

I’ve been a Scientific American subscriber since I was 12 years old, and it helped shape the way I look at the world. SciAm always educates and delights me, and inspires a sense of awe for our vast, beautiful universe. I hope it does that for you, too.

If you subscribe to Scientific American, you help ensure that our coverage is centered on meaningful research and discovery; that we have the resources to report on the decisions that threaten labs across the U.S.; and that we support both budding and working scientists at a time when the value of science itself too often goes unrecognized.

In return, you get essential news, captivating podcasts, brilliant infographics, can't-miss newsletters, must-watch videos, challenging games, and the science world's best writing and reporting. You can even gift someone a subscription.

There has never been a more important time for us to stand up and show why science matters. I hope you’ll support us in that mission.

Thank you,

David M. Ewalt, Editor in Chief, Scientific American

Subscribe