PhysReason: A Comprehensive Benchmark towards Physics-Based Reasoning
Zhang, Xinyu, Dong, Yuxuan, Wu, Yanrui, Huang, Jiaxing, Jia, Chengyou, Fernando, Basura, Shou, Mike Zheng, Zhang, Lingling, Liu, Jun
–arXiv.org Artificial Intelligence
Large language models demonstrate remarkable capabilities across various domains, especially mathematics and logic reasoning. However, current evaluations overlook physics-based reasoning - a complex task requiring physics theorems and constraints. We present PhysReason, a 1,200-problem benchmark comprising knowledge-based (25%) and reasoning-based (75%) problems, where the latter are divided into three difficulty levels (easy, medium, hard). Notably, problems require an average of 8.1 solution steps, with hard requiring 15.6, reflecting the complexity of physics-based reasoning. We propose the Physics Solution Auto Scoring Framework, incorporating efficient answer-level and comprehensive step-level evaluations. Top-performing models like Deepseek-R1, Gemini-2.0-Flash-Thinking, and o3-mini-high achieve less than 60% on answer-level evaluation, with performance dropping from knowledge questions (75.11%) to hard problems (31.95%). Through step-level evaluation, we identified four key bottlenecks: Physics Theorem Application, Physics Process Understanding, Calculation, and Physics Condition Analysis. These findings position PhysReason as a novel and comprehensive benchmark for evaluating physics-based reasoning capabilities in large language models. Our code and data will be published at https:/dxzxy12138.github.io/PhysReason.
arXiv.org Artificial Intelligence
Feb-17-2025
- Country:
- Asia (0.93)
- Genre:
- Research Report (1.00)
- Industry:
- Education
- Assessment & Standards (1.00)
- Educational Setting (0.67)
- Education
- Technology: