Source2Synth: Synthetic Data Generation and Curation Grounded in Real Data Sources
Lupidi, Alisia, Gemmell, Carlos, Cancedda, Nicola, Dwivedi-Yu, Jane, Weston, Jason, Foerster, Jakob, Raileanu, Roberta, Lomeli, Maria
–arXiv.org Artificial Intelligence
Large Language Models still struggle in challenging scenarios that leverage structured data, complex reasoning, or tool usage. In this paper, we propose Source2Synth: a new method that can be used for teaching LLMs new skills without relying on costly human annotations. Source2Synth takes as input a custom data source and produces synthetic data points with intermediate reasoning steps grounded in real-world sources. Source2Synth improves the dataset quality by discarding low-quality generations based on their answerability. We demonstrate the generality of this approach by applying it to two challenging domains: we test reasoning abilities in multi-hop question answering (MHQA), and tool usage in tabular question answering (TQA). Our method improves performance by 25.51% for TQA on WikiSQL and 22.57% for MHQA on HotPotQA compared to the fine-tuned baselines.
arXiv.org Artificial Intelligence
Sep-12-2024
- Country:
- Asia
- China > Beijing
- Beijing (0.04)
- Middle East
- Saudi Arabia > Asir Province
- Abha (0.04)
- UAE (0.04)
- Saudi Arabia > Asir Province
- China > Beijing
- Europe > United Kingdom
- England > Oxfordshire > Oxford (0.04)
- North America
- Canada (0.04)
- Mexico (0.04)
- United States (0.95)
- Asia
- Genre:
- Research Report (0.50)
- Industry:
- Technology: