iREL at SemEval-2024 Task 9: Improving Conventional Prompting Methods for Brain Teasers
Gupta, Harshit, Chaudhary, Manav, Raha, Tathagata, Subramanian, Shivansh, Varma, Vasudeva
–arXiv.org Artificial Intelligence
This paper describes our approach for SemEval-2024 Task 9: BRAINTEASER: A Novel Task Defying Common Sense. The BRAINTEASER task comprises multiple-choice Question Answering designed to evaluate the models' lateral thinking capabilities. It consists of Sentence Puzzle and Word Puzzle subtasks that require models to defy default common-sense associations and exhibit unconventional thinking. We propose a unique strategy to improve the performance of pre-trained language models, notably the Gemini 1.0 Pro Model, in both subtasks. We employ static and dynamic few-shot prompting techniques and introduce a model-generated reasoning strategy that utilizes the LLM's reasoning capabilities to improve performance. Our approach demonstrated significant improvements, showing that it performed better than the baseline models by a considerable margin but fell short of performing as well as the human annotators, thus highlighting the efficacy of the proposed strategies.
arXiv.org Artificial Intelligence
May-25-2024
- Country:
- Asia > Singapore (0.04)
- North America > Mexico
- Mexico City > Mexico City (0.04)
- Genre:
- Research Report > New Finding (0.48)
- Industry:
- Education (0.49)
- Technology: