ToolMind Technical Report: A Large-Scale, Reasoning-Enhanced Tool-Use Dataset
Yang, Chen, Le, Ran, Xing, Yun, An, Zhenwei, Chen, Zongchao, Zhao, Wayne Xin, Song, Yang, Zhang, Tao
–arXiv.org Artificial Intelligence
Large Language Model (LLM) agents have developed rapidly in recent years to solve complex real-world problems using external tools. However, the scarcity of high-quality trajectories still hinders the development of stronger LLM agents. Most existing works on multi-turn dialogue synthesis validate correctness only at the trajectory level, which may overlook turn-level errors that can propagate during training and degrade model performance. To address these limitations, we introduce ToolMind, a large-scale, high-quality tool-agentic dataset with 160k synthetic data instances generated using over 20k tools and 200k augmented open-source data instances. Our data synthesis pipeline first constructs a function graph based on parameter correlations and then uses a multi-agent framework to simulate realistic user-assistant-tool interactions. Beyond trajectory-level validation, we employ fine-grained turn-level filtering to remove erroneous or suboptimal steps, ensuring that only high-quality reasoning traces are retained. This approach mitigates error amplification during training while preserving self-corrective reasoning signals essential for robust tool-use learning. Models fine-tuned on ToolMind show significant improvements over baselines on several benchmarks.
arXiv.org Artificial Intelligence
Dec-8-2025
- Country:
- Genre:
- Research Report (0.82)
- Technology: