DeepMind’s AI Surpasses Math Olympiad Champions in Geometry, Paving Way for Hybrid AI Breakthroughs

Marian Schultz
February 8, 2025
Become smarter in just 5 minutes
Feeeed serve up the most interesting stories in AI, tech, business, science, and more—all in 5 minutes or less.
Booyah! The first newsletter will go out early next week.
Oops! Something went wrong while submitting the form.

Why it matters:

AlphaGeometry2 solves complex geometry problems better than top-tier math competitors, signaling progress in AI’s ability to mimic human-like reasoning — a hurdle for chatbots like ChatGPT.

How it works

  • Hybrid brain: Marries Google’s Gemini AI (neural network) for creative “ideas” with a rules-based engine to check logical steps.
  • Synthetic data: Trained on 300M+ fake theorems to bypass scarce real-world math proofs.
  • Test results: Nailed 42/50 past Olympiad problems, edging past gold medalists’ average of 40.9. But stumbled on 20/29 tougher, unseen puzzles.

Reality check

  • Flaws: Chokes on problems with shifting points or nonlinear equations.
  • Not pure AI: Relies on hand-coded rules — not just learned intuition.

Between the lines: This isn’t just about math. Hybrid models (neural + symbolic systems) could help AI tackle engineering or physics tasks requiring precise proofs — think verifying bridge designs or drug molecules.

The catch: AI still can’t reliably ace basic logic puzzles, notes Carnegie Mellon’s Vince Conitzer: “It’s striking [these systems] solve Olympiad problems but fumble simple things. We urgently need to understand their limits as they scale.”

What’s next: DeepMind plans to expand the tech to broader math/science fields — but fully self-sufficient AI reasoning remains years away.

🔍 Go deeper: AlphaGeometry2 study | IMO problems sample*

Stay smarter on AI, Tech, Business and What Matters in 5 minutes

Get the 5-minute news brief to keep you in loop. Free forever.