OR-Toolformer: Modeling and Solving Operations Research Problems with Tool Augmented Large Language Models
Zhang, Jianzhang, Zhou, Jialong, Liu, Chuang
–arXiv.org Artificial Intelligence
Large language models (LLMs) demonstrate strong mathematical reasoning, but reliance on closed-source APIs for OR tasks raises privacy concerns, and training open-source models from scratch incurs high compute costs. We introduce OR-Toolformer, which fine-tunes Llama-3.1-8B-Instruct with a semi-automatic data synthesis pipeline that generates diverse OR problem-answer pairs and augments the model with external solvers to produce API calls. On three of four standard benchmarks, OR-Toolformer achieves up to 80.1% execution accuracy, exceeding size-matched baselines by over 4.3%. In zero-shot evaluation on two unseen OR problem types, it attains 54% average accuracy, a 21 percentage-point improvement over the strongest baseline. These findings validate the efficacy of tool-augmented fine-tuning LLMs for accurate and generalizable OR problem modeling and solving.
arXiv.org Artificial Intelligence
Oct-3-2025
- Country:
- Asia
- China > Zhejiang Province
- Hangzhou (0.04)
- Middle East > Jordan (0.04)
- Myanmar > Tanintharyi Region
- Dawei (0.05)
- Thailand > Bangkok
- Bangkok (0.05)
- China > Zhejiang Province
- North America > United States
- Florida > Miami-Dade County > Miami (0.05)
- Asia
- Genre:
- Research Report (0.64)
- Industry:
- Information Technology > Security & Privacy (0.67)
- Technology: