Towards Better Understanding Table Instruction Tuning: Decoupling the Effects from Data versus Models
Deng, Naihao, Zhang, Sheng, Zhu, Henghui, Chang, Shuaichen, Zhang, Jiani, Li, Alexander Hanbo, Hang, Chung-Wei, Kobayashi, Hideo, Hu, Yiqun, Ng, Patrick
–arXiv.org Artificial Intelligence
Recent advances in natural language processing have leveraged instruction tuning to enhance Large Language Models (LLMs) for table-related tasks. However, previous works train different base models with different training data, lacking an apples-to-apples comparison across the result table LLMs. To address this, we fine-tune base models from the Mistral, OLMo, and Phi families on existing public training datasets. Our replication achieves performance on par with or surpassing existing table LLMs, establishing new state-of-the-art performance on Hitab, a table question-answering dataset. More importantly, through systematic out-of-domain evaluation, we decouple the contributions of training data and the base model, providing insight into their individual impacts. In addition, we assess the effects of table-specific instruction tuning on general-purpose benchmarks, revealing trade-offs between specialization and generalization.
arXiv.org Artificial Intelligence
Jan-24-2025
- Country:
- Asia
- China > Beijing
- Beijing (0.04)
- Middle East
- Iraq > Erbil Governorate
- Erbil (0.04)
- UAE > Abu Dhabi Emirate
- Abu Dhabi (0.04)
- Iraq > Erbil Governorate
- Thailand > Bangkok
- Bangkok (0.04)
- China > Beijing
- Europe
- Ireland > Leinster
- County Dublin > Dublin (0.04)
- Netherlands (0.04)
- Switzerland > Geneva
- Geneva (0.04)
- United Kingdom (0.04)
- Ireland > Leinster
- North America
- Dominican Republic (0.04)
- Mexico > Mexico City
- Mexico City (0.04)
- United States
- Michigan (0.04)
- New York (0.04)
- South Carolina (0.04)
- Texas > Travis County
- Austin (0.04)
- Asia
- Genre:
- Research Report > New Finding (0.68)
- Industry:
- Government (0.46)
- Leisure & Entertainment > Sports (0.46)
- Technology: