Gemini Deep Think Wins Gold at IMO with Parallel Reasoning

AI vs. the IMO: A New Benchmark in Mathematical Reasoning
At the 2025 International Math Olympiad (IMO) in Perth, Australia, Google’s latest AI model, Gemini Deep Think, matched wits with the world’s top pre-university mathematicians—and came away with a gold medal. DeepMind collaborated directly with IMO organizers to ensure the model was evaluated under exactly the same conditions as human participants: six proof-based problems, a strict 4.5-hour time limit, and full requirement to show every step of the solution.
From Silver to Gold: Evolution of DeepMind’s IMO Strategy
Last year, a hybrid system built on AlphaProof and AlphaGeometry 2 secured a silver medal by correctly solving four out of six questions. For 2025, DeepMind introduced Gemini Deep Think, a model architected for parallel simulated reasoning rather than a single linear chain of thought. According to Thang Luong, DeepMind senior scientist and head of the IMO team, “Deep Think was trained end-to-end on natural language problems—no manual translation to domain-specific code was needed. It reasons in multiple channels simultaneously, then aggregates proofs into one coherent answer.”
Technical Architecture of Gemini Deep Think
Under the hood, Gemini Deep Think is a 110 billion-parameter transformer ensemble running on Google’s TPU v5 Pods. Key architectural enhancements include:
- Mixture-of-Experts (MoE) layers that dynamically activate only 20% of the network per problem, reducing compute cost by 40% while preserving performance.
- Multi-Chain-of-Thought modules that spawn 8 parallel reasoning traces, each exploring different proof strategies (e.g., algebraic, combinatorial, geometric), then perform cross-attention fusion to select the tightest solution.
- Extended Context Window of 128,000 tokens to maintain all intermediate lemmas and sub-proofs in memory.
Reinforcement Learning with Long-Form Mathematical Feedback
Rather than relying solely on final-answer supervision, DeepMind designed a two-stage reinforcement learning pipeline:
- Proof-Level Reward Model: Trained on 10,000 graded IMO-style solutions, scoring each proof step for correctness, elegance, and self-containment.
- Curriculum Learning: Beginning with single-step algebraic problems, the model gradually progressed to multi-disciplinary proofs, effectively “bootstrapping” its capability across complexity levels.
“This approach yields robust, transparent reasoning chains,” said Luong. “Our ablations show Deep Think’s token-level perplexity on proof text dropped by 30% compared to last year’s model.”
Comparative Performance and Latest Industry Developments
Among AI contenders at IMO 2025:
- Gemini Deep Think (DeepMind): 5/6 correct → Gold
- GPT-4o (OpenAI): 4/6 correct → Unofficial silver, pending IMO certification
- LLaMA 3 (Meta): 4/6 correct → Silver
- Minerva II (Google Research): 3/6 correct → Honorable mention
In June 2025, OpenAI announced GPT-4.5 with improved mathematical reasoning, though it opted to use an external panel of ex-IMO participants for grading rather than the official IMO process. By contrast, DeepMind confirmed with the IMO jury that Deep Think adhered rigorously to all rules and presentation standards.
Implications for Formal Verification and Educational Tools
Deep Think’s capacity to generate detailed, step-by-step proofs has immediate applications in formal verification frameworks like Coq and Lean. Researchers at MIT have already begun experiments integrating Deep Think outputs into automated theorem-proving pipelines, reducing proof-development time by up to 50%. In education, platforms such as Khan Academy and Coursera are in talks with DeepMind to embed similar models for real-time homework feedback, potentially transforming how advanced mathematics is taught and learned.
Future Directions and Next Year’s Goals
- Perfect Score Pursuit: DeepMind aims to achieve 6/6 correct by integrating symbolic reasoning modules with its neural backbone.
- Open Benchmarking: Publishing an academic paper and releasing test harnesses for the community under an open-source license.
- Enterprise Rollout: Deploying the IMO-tuned model within Google AI Ultra (subscription $250/month) for mathematicians, R&D labs, and advanced analytics teams.
As the bar for machine reasoning continues to rise, Gemini Deep Think’s gold-medal performance marks not just a milestone in AI capabilities, but a harbinger of deeper collaborations between human and machine intellect.