From Babble to Words: Pre-Training Language Models on Continuous Streams of Phonemes
Goriely, Zébulon, Martinez, Richard Diehl, Caines, Andrew, Beinborn, Lisa, Buttery, Paula
–arXiv.org Artificial Intelligence
Language models are typically trained on large corpora of text in their default orthographic form. However, this is not the only option; representing data as streams of phonemes can offer unique advantages, from deeper insights into phonological language acquisition to improved performance on sound-based tasks. The challenge lies in evaluating the impact of phoneme-based training, as most benchmarks are also orthographic. To address this, we develop a pipeline to convert text datasets into a continuous stream of phonemes. We apply this pipeline to the 100-million-word pre-training dataset from the BabyLM challenge, as well as to standard language and grammatical benchmarks, enabling us to pre-train and evaluate a model using phonemic input representations. Our results show that while phoneme-based training slightly reduces performance on traditional language understanding tasks, it offers valuable analytical and practical benefits.
arXiv.org Artificial Intelligence
Oct-30-2024
- Country:
- Asia
- Europe
- Belgium > Brussels-Capital Region
- Brussels (0.04)
- Bulgaria (0.04)
- Croatia (0.04)
- Germany
- Berlin (0.04)
- Lower Saxony > Gottingen (0.14)
- Ireland > Leinster
- County Dublin > Dublin (0.04)
- Italy (0.04)
- Spain > Catalonia
- Barcelona Province > Barcelona (0.04)
- United Kingdom > England
- Cambridgeshire > Cambridge (0.28)
- Belgium > Brussels-Capital Region
- North America
- Canada > Ontario
- Toronto (0.04)
- Mexico > Mexico City
- Mexico City (0.04)
- United States
- District of Columbia > Washington (0.04)
- Minnesota > Hennepin County
- Minneapolis (0.14)
- Washington > King County
- Seattle (0.04)
- Canada > Ontario
- Oceania > Australia (0.04)
- Genre:
- Research Report > New Finding (0.86)
- Industry:
- Information Technology (0.68)
- Technology: