MIA 2022 Shared Task Submission: Leveraging Entity Representations, Dense-Sparse Hybrids, and Fusion-in-Decoder for Cross-Lingual Question Answering
Tu, Zhucheng, Padmanabhan, Sarguna Janani
–arXiv.org Artificial Intelligence
We describe our two-stage system for the Multilingual Information Access (MIA) 2022 Shared Task on Cross-Lingual Open-Retrieval Question Answering. The first stage consists of multilingual passage retrieval with a hybrid dense and sparse retrieval strategy. The second stage consists of a reader which outputs the answer from the top passages returned by the first stage. We show the efficacy of using a multilingual language model with entity representations in pretraining, sparse retrieval signals to help dense retrieval, and Fusion-in-Decoder. On the development set, we obtain 43.46 F1 on XOR-TyDi QA and 21.99 F1 on MKQA, for an average F1 score of 32.73. On the test set, we obtain 40.93 F1 on XOR-TyDi QA and 22.29 F1 on MKQA, for an average F1 score of 31.61. We improve over the official baseline by over 4 F1 points on both the development and test sets.
arXiv.org Artificial Intelligence
Jul-18-2022
- Country:
- Europe (0.04)
- North America > United States
- Utah (0.04)
- Genre:
- Research Report > New Finding (0.34)
- Industry:
- Government (0.70)
- Leisure & Entertainment > Sports
- Soccer (0.94)
- Technology: