Logically Consistent Language Models via Neuro-Symbolic Integration
Calanzone, Diego, Teso, Stefano, Vergari, Antonio
–arXiv.org Artificial Intelligence
Large language models (LLMs) are a promising venue for natural language understanding and generation. However, current LLMs are far from reliable: they are prone to generating non-factual information and, more crucially, to contradicting themselves when prompted to reason about relations between entities of the world. These problems are currently addressed with large scale fine-tuning or by delegating reasoning to external tools. In this work, we strive for a middle ground and introduce a loss based on neuro-symbolic reasoning that teaches an LLM to be logically consistent with an external set of facts and rules and improves self-consistency even when the LLM is fine-tuned on a limited set of facts. Our approach also allows to easily combine multiple logical constraints at once in a principled way, delivering LLMs that are more consistent w.r.t. all constraints and improve over several baselines w.r.t. a given constraint. Moreover, our method allows LLMs to extrapolate to unseen but semantically similar factual knowledge, represented in unseen datasets, more systematically.
arXiv.org Artificial Intelligence
Sep-9-2024
- Country:
- Africa > Ethiopia
- Addis Ababa > Addis Ababa (0.04)
- Asia > Middle East
- Republic of Türkiye (0.04)
- North America > United States
- Massachusetts > Suffolk County > Boston (0.04)
- Africa > Ethiopia
- Genre:
- Research Report > New Finding (0.46)
- Technology: