TTPA: Token-level Tool-use Preference Alignment Training Framework with Fine-grained Evaluation
Huang, Chengrui, Gao, Shen, Shi, Zhengliang, Wang, Dongsheng, Shang, Shuo
–arXiv.org Artificial Intelligence
Existing tool-learning methods usually rely on supervised fine-tuning, they often overlook fine-grained optimization of internal tool call details, leading to limitations in preference alignment and error discrimination. To overcome these challenges, we propose Token-level Tool-use Preference Alignment Training Framework (TTPA), a training paradigm for constructing token-level tool-use preference datasets that align LLMs with fine-grained preferences using a novel error-oriented scoring mechanism. TTPA first introduces reversed dataset construction, a method for creating high-quality, multi-turn tool-use datasets by reversing the generation flow. Additionally, we propose Token-level Preference Sampling (TPS) to capture fine-grained preferences by modeling token-level differences during generation. To address biases in scoring, we introduce the Error-oriented Scoring Mechanism (ESM), which quantifies tool-call errors and can be used as a training signal. Extensive experiments on three diverse benchmark datasets demonstrate that TTPA significantly improves tool-using performance while showing strong generalization ability across models and datasets.
arXiv.org Artificial Intelligence
May-27-2025
- Country:
- Asia
- China (0.04)
- Myanmar > Tanintharyi Region
- Dawei (0.04)
- Asia
- Genre:
- Research Report (0.82)
- Technology: