Language Models Can Improve Event Prediction by Few-Shot Abductive Reasoning
Shi, Xiaoming, Xue, Siqiao, Wang, Kangrui, Zhou, Fan, Zhang, James Y., Zhou, Jun, Tan, Chenhao, Mei, Hongyuan
–arXiv.org Artificial Intelligence
Large language models have shown astonishing performance on a wide range of reasoning tasks. In this paper, we investigate whether they could reason about real-world events and help improve the prediction performance of event sequence models. We design LAMP, a framework that integrates a large language model in event prediction. Particularly, the language model performs abductive reasoning to assist an event sequence model: the event model proposes predictions on future events given the past; instructed by a few expert-annotated demonstrations, the language model learns to suggest possible causes for each proposal; a search module finds out the previous events that match the causes; a scoring function learns to examine whether the retrieved events could actually cause the proposal. Through extensive experiments on several challenging real-world datasets, we demonstrate that our framework -- thanks to the reasoning capabilities of large language models -- could significantly outperform the state-of-the-art event sequence models.
arXiv.org Artificial Intelligence
Oct-7-2023
- Country:
- Asia (0.93)
- Europe (0.93)
- North America > United States (0.46)
- Genre:
- Research Report (1.00)
- Industry:
- Energy > Oil & Gas (0.67)
- Government (0.93)
- Health & Medicine (1.00)
- Technology: