Goto

Collaborating Authors

 imo problem


Gold-medalist Performance in Solving Olympiad Geometry with AlphaGeometry2

Chervonyi, Yuri, Trinh, Trieu H., Olšák, Miroslav, Yang, Xiaomeng, Nguyen, Hoang, Menegali, Marcelo, Jung, Junehyuk, Verma, Vikas, Le, Quoc V., Luong, Thang

arXiv.org Artificial Intelligence

We present AlphaGeometry2, a significantly improved version of AlphaGeometry introduced in Trinh et al. (2024), which has now surpassed an average gold medalist in solving Olympiad geometry problems. To achieve this, we first extend the original AlphaGeometry language to tackle harder problems involving movements of objects, and problems containing linear equations of angles, ratios, and distances. This, together with other additions, has markedly improved the coverage rate of the AlphaGeometry language on International Math Olympiads (IMO) 2000-2024 geometry problems from 66% to 88%. The search process of AlphaGeometry2 has also been greatly improved through the use of Gemini architecture for better language modeling, and a novel knowledge-sharing mechanism that combines multiple search trees. Together with further enhancements to the symbolic engine and synthetic data generation, we have significantly boosted the overall solving rate of AlphaGeometry2 to 84% for $\textit{all}$ geometry problems over the last 25 years, compared to 54% previously. AlphaGeometry2 was also part of the system that achieved silver-medal standard at IMO 2024 https://dpmd.ai/imo-silver. Last but not least, we report progress towards using AlphaGeometry2 as a part of a fully automated system that reliably solves geometry problems directly from natural language input.


A Lean Dataset for International Math Olympiad: Small Steps towards Writing Math Proofs for Hard Problems

Yousefzadeh, Roozbeh, Cao, Xuenan

arXiv.org Artificial Intelligence

Using AI to write formal proofs for mathematical problems is a challenging task that has seen some advancements in recent years. Automated systems such as Lean can verify the correctness of proofs written in formal language, yet writing the proofs in formal language can be challenging for humans and machines. The miniF2F benchmark has 20 IMO problems in its testing set, yet formal proofs are available only for 7 of these problems (3 of which are written only by mathematicians). The model with best accuracy can only prove 4 of these 20 IMO problems, from 1950s and 60s, while its training set is a secret. In this work, we write complete, original formal proofs for the remaining 13 IMO problems in Lean along with 3 extra problems from IMO 2022 and 2023. This effort expands the availability of proof currently in the public domain by creating 5,150 lines of Lean proof. The goal of the paper is to pave the way for developing AI models that can automatically write the formal proofs for all the IMO problems in miniF2F and beyond. In this pursuit, we devise a method to decompose the proof of these problems into their building blocks, constructing a dataset of about 900 lemmas with 25,500 lines of Lean code. These lemmas are not trivial, yet they are approachable, providing the opportunity to evaluate and diagnose the failures and successes of AI models. We then evaluate the ability of GPT-4 in writing formal proofs for these lemmas with zero shot prompting, CoT reasoning and lemma retrieval. In evaluating the responses, we also analyze the confounding factor of LLM's ability to write the proofs in natural language vs Lean language.


FIMO: A Challenge Formal Dataset for Automated Theorem Proving

Liu, Chengwu, Shen, Jianhao, Xin, Huajian, Liu, Zhengying, Yuan, Ye, Wang, Haiming, Ju, Wei, Zheng, Chuanyang, Yin, Yichun, Li, Lin, Zhang, Ming, Liu, Qun

arXiv.org Artificial Intelligence

We present FIMO, an innovative dataset comprising formal mathematical problem statements sourced from the International Mathematical Olympiad (IMO) Shortlisted Problems. Designed to facilitate advanced automated theorem proving at the IMO level, FIMO is currently tailored for the Lean formal language. It comprises 149 formal problem statements, accompanied by both informal problem descriptions and their corresponding LaTeX-based informal proofs. Through initial experiments involving GPT-4, our findings underscore the existing limitations in current methodologies, indicating a substantial journey ahead before achieving satisfactory IMO-level automated theorem proving outcomes.


At the Math Olympiad, Computers Prepare to Go for the Gold

#artificialintelligence

The 61st International Mathematical Olympiad, or IMO, begins today. It may go down in history for at least two reasons: Due to the COVID-19 pandemic it's the first time the event has been held remotely, and it may also be the last time that artificial intelligence doesn't compete. Indeed, researchers view the IMO as the ideal proving ground for machines designed to think like humans. If an AI system can excel here, it will have matched an important dimension of human cognition. "The IMO, to me, represents the hardest class of problems that smart people can be taught to solve somewhat reliably," said Daniel Selsam of Microsoft Research.


Can Artificial Intelligence Win Olympics ?

#artificialintelligence

Data scientists are trying to build an AI system that can win a gold medal at the world's premier math competition Indeed, researchers view the IMO as the ideal proving ground for machines designed to think like humans. If an AI system can excel here, it will have matched an important dimension of human cognition. "The IMO, to me, represents the hardest class of problems that smart people can be taught to solve somewhat reliably," said Daniel Selsam of Microsoft Research. Selsam is a founder of the IMO Grand Challenge, whose goal is to train an AI system to win a gold medal at the world's premier math competition. The International Mathematical Olympiad (IMO) is a mathematical olympiad for pre-college students, and is the oldest of the International Science Olympiads.


At the Math Olympiad, Computers Prepare to Go for the Gold - Facts So Romantic

Nautilus

Reprinted with permission from Quanta Magazine's Abstractions blog. The 61st International Mathematical Olympiad, or IMO, began yesterday. It may go down in history for at least two reasons: Due to the COVID-19 pandemic it's the first time the event has been held remotely, and it may also be the last time that artificial intelligence doesn't compete. Indeed, researchers view the IMO as the ideal proving ground for machines designed to think like humans. If an AI system can excel here, it will have matched an important dimension of human cognition.

  Country: North America > Cuba (0.05)
  Industry:
  Technology: Information Technology > Artificial Intelligence > Games > Go (0.40)