Latent Compositional Representations Improve Systematic Generalization in Grounded Question Answering
Bogin, Ben, Subramanian, Sanjay, Gardner, Matt, Berant, Jonathan
–arXiv.org Artificial Intelligence
Answering questions that involve multi-step reasoning requires decomposing them and using the answers of intermediate steps to reach the final answer. However, state-of-the-art models in grounded question answering often do not explicitly perform decomposition, leading to difficulties in generalization to out-of-distribution examples. In this work, we propose a model that computes a representation and denotation for all question spans in a bottom-up, compositional manner using a CKY-style parser. Our model effectively induces latent trees, driven by end-to-end (the answer) supervision only. We show that this inductive bias towards tree structures dramatically improves systematic generalization to out-of-distribution examples compared to strong baselines on an arithmetic expressions benchmark as well as on CLOSURE, a dataset that focuses on systematic generalization of models for grounded question answering. On this challenging dataset, our model reaches an accuracy of 92.8%, significantly higher than prior models that almost perfectly solve the task on a random, in-distribution split.
arXiv.org Artificial Intelligence
Jul-1-2020
- Country:
- Asia (0.93)
- Europe (0.68)
- North America > United States
- Minnesota (0.28)
- Genre:
- Research Report (0.70)
- Technology: