kaylee
Exploring Next Token Prediction in Theory of Mind (ToM) Tasks: Comparative Experiments with GPT-2 and LLaMA-2 AI Models
Yadav, Pavan, Khandalkar, Nikhil, Shinde, Krishna, Ramegowda, Lokesh B., Das, Rajarshi
Language models have made significant progress in generating coherent text and predicting next tokens based on input prompts. This study compares the next-token prediction performance of two well-known models: OpenAI's GPT-2 and Meta's Llama-2-7b-chat-hf on Theory of Mind (ToM) tasks. To evaluate their capabilities, we built a dataset from 10 short stories sourced from the Explore ToM Dataset. We enhanced these stories by programmatically inserting additional sentences (infills) using GPT-4, creating variations that introduce different levels of contextual complexity. This setup enables analysis of how increasing context affects model performance. We tested both models under four temperature settings (0.01, 0.5, 1.0, 2.0) and evaluated their ability to predict the next token across three reasoning levels. Zero-order reasoning involves tracking the state, either current (ground truth) or past (memory). First-order reasoning concerns understanding another's mental state (e.g., "Does Anne know the apple is salted?"). Second-order reasoning adds recursion (e.g., "Does Anne think that Charles knows the apple is salted?"). Our results show that adding more infill sentences slightly reduces prediction accuracy, as added context increases complexity and ambiguity. Llama-2 consistently outperforms GPT-2 in prediction accuracy, especially at lower temperatures, demonstrating greater confidence in selecting the most probable token. As reasoning complexity rises, model responses diverge more. Notably, GPT-2 and Llama-2 display greater variability in predictions during first- and second-order reasoning tasks. These findings illustrate how model architecture, temperature, and contextual complexity influence next-token prediction, contributing to a better understanding of the strengths and limitations of current language models.
- North America > United States (0.14)
- Asia > India > West Bengal > Kolkata (0.05)
- Asia > India > Karnataka > Bengaluru (0.04)
What Happens When Two Artificial Intelligences Try To Prank Each Other?
Our Artificial Intelligence app, Hugging Face, has been running smoothly following a big influx of new users. It's a normal day, and I'm looking over activity readouts when suddenly the app grinds to a complete halt. Thousands of teens chatting with their AI friends are getting nothing but silence in return. I pull up Slack, and ask the tech team if we are down. Julien, my co-founder and the CTO of Hugging Face, looks over the brains of our AIs, and comes up with nothing out of the ordinary.