natural language problem
- Research Report (0.68)
- Instructional Material > Course Syllabus & Notes (0.50)
- Research Report (0.68)
- Instructional Material > Course Syllabus & Notes (0.50)
LEAN-GitHub: Compiling GitHub LEAN repositories for a versatile LEAN prover
Wu, Zijian, Wang, Jiayu, Lin, Dahua, Chen, Kai
Recently, large language models have presented promising results in aiding formal mathematical reasoning. However, their performance is restricted due to the scarcity of formal theorem-proving data, which requires additional effort to be extracted from raw formal language corpora. Meanwhile, a significant amount of human-written formal language corpora remains underutilized. To address this issue, we propose LEAN-GitHub, a dataset consisting of large-scale formal data extracted from almost all Lean 4 repositories on GitHub. After fine-tuning InternLM-math-plus on this dataset, our model achieved accuracies of 48.8% with a single pass and 54.5% with 64 passes on the Lean 4 miniF2F test, surpassing state-of-the-art method at 52%. And it also achieves state-of-the-art on two other Lean 4 benchmarks (ProofNet and Putnam) targeting different fields/levels of math. These results demonstrate that our proposed dataset is beneficial for formal reasoning on a wide range of math topics. We open-source our model at https://GitHub. com/InternLM/InternLM-Math and our data at https://huggingface.co/ datasets/InternLM/Lean-GitHub
- Asia > China > Shanghai > Shanghai (0.04)
- North America > United States > New York > New York County > New York City (0.04)
- Europe > Germany > Berlin (0.04)
- Asia > China > Hong Kong (0.04)
- Research Report > Promising Solution (0.48)
- Research Report > New Finding (0.34)
Lean Workbook: A large-scale Lean problem set formalized from natural language math problems
Ying, Huaiyuan, Wu, Zijian, Geng, Yihan, Wang, Jiayu, Lin, Dahua, Chen, Kai
Large language models have demonstrated impressive capabilities across various natural language processing tasks, especially in solving mathematical problems. However, large language models are not good at math theorem proving using formal languages like Lean. A significant challenge in this area is the scarcity of training data available in these formal languages. To address this issue, we propose a novel pipeline that iteratively generates and filters synthetic data to translate natural language mathematical problems into Lean 4 statements, and vice versa. Our results indicate that the synthetic data pipeline can provide useful training data and improve the performance of LLMs in translating and understanding complex mathematical problems and proofs. Our final dataset contains about 57K formal-informal question pairs along with searched proof from the math contest forum and 21 new IMO questions.
- Instructional Material > Course Syllabus & Notes (0.50)
- Research Report > New Finding (0.34)