An Autoencoder Approach to Learning Bilingual Word Representations Sarath Chandar A P
–Neural Information Processing Systems
Cross-language learning allows one to use training data from one language to build models for a different language. Many approaches to bilingual learning require that we have word-level alignment of sentences from parallel corpora. In this work we explore the use of autoencoder-based methods for cross-language learning of vectorial word representations that are coherent between two languages, while not relying on word-level alignments. We show that by simply learning to reconstruct the bag-of-words representations of aligned sentences, within and between languages, we can in fact learn high-quality representations and do without word alignments. We empirically investigate the success of our approach on the problem of cross-language text classification, where a classifier trained on a given language (e.g., English) must learn to generalize to a different language (e.g., German). In experiments on 3 language pairs, we show that our approach achieves state-of-the-art performance, outperforming a method exploiting word alignments and a strong machine translation baseline.
Neural Information Processing Systems
Mar-13-2024, 07:15:42 GMT
- Country:
- Africa > Middle East
- Egypt > Giza Governorate > Giza (0.04)
- Asia
- Europe
- Bulgaria > Sofia City Province
- Sofia (0.04)
- Czechia > Prague (0.04)
- Germany > Saarland (0.04)
- Sweden > Vaestra Goetaland
- Gothenburg (0.04)
- Bulgaria > Sofia City Province
- North America > United States
- Maryland > Baltimore (0.14)
- Oregon > Multnomah County
- Portland (0.04)
- Pennsylvania (0.04)
- Africa > Middle East
- Industry:
- Education > Curriculum > Subject-Specific Education (0.45)
- Technology: