A Deep Reinforcement Learning based Approach to Learning Transferable Proof Guidance Strategies
Crouse, Maxwell, Whitehead, Spencer, Abdelaziz, Ibrahim, Makni, Bassem, Cornelio, Cristina, Kapanipathi, Pavan, Pell, Edwin, Srinivas, Kavitha, Thost, Veronika, Witbrock, Michael, Fokoue, Achille
–arXiv.org Artificial Intelligence
Traditional first-order logic (FOL) reasoning systems usually rely on manual heuristics for proof guidance. We propose TRAIL: a system that learns to perform proof guidance using reinforcement learning. A key design principle of our system is that it is general enough to allow transfer to problems in different domains that do not share the same vocabulary of the training set. To do so, we developed a novel representation of the internal state of a prover in terms of clauses and inference actions, and a novel neural-based attention mechanism to learn interactions between clauses. We demonstrate that this approach enables the system to generalize from training to test data across domains with different vocabularies, suggesting that the neural architecture in TRAIL is well suited for representing and processing of logical formalisms.
arXiv.org Artificial Intelligence
Nov-5-2019
- Country:
- Africa > Botswana
- North-West District > Maun (0.04)
- Europe
- Belgium > Brussels-Capital Region
- Brussels (0.04)
- Portugal > Lisbon
- Lisbon (0.04)
- Switzerland > Zürich
- Zürich (0.04)
- Belgium > Brussels-Capital Region
- North America > United States
- Illinois (0.04)
- Massachusetts > Middlesex County
- Cambridge (0.04)
- Oceania > New Zealand
- North Island > Auckland Region > Auckland (0.04)
- Africa > Botswana
- Genre:
- Research Report (0.82)
- Technology: