capital city
Sparse Feature Coactivation Reveals Causal Semantic Modules in Large Language Models
Deng, Ruixuan, Hu, Xiaoyang, Gilberti, Miles, Storks, Shane, Taxali, Aman, Angstadt, Mike, Sripada, Chandra, Chai, Joyce
We identify semantically coherent, context-consistent network components in large language models (LLMs) using coactivation of sparse autoencoder (SAE) features collected from just a handful of prompts. Focusing on concept-relation prediction tasks, we show that ablating these components for concepts (e.g., countries and words) and relations (e.g., capital city and translation language) changes model outputs in predictable ways, while amplifying these components induces counterfactual responses. Notably, composing relation and concept components yields compound counterfactual outputs. Further analysis reveals that while most concept components emerge from the very first layer, more abstract relation components are concentrated in later layers. Lastly, we show that extracted components more comprehensively capture concepts and relations than individual features while maintaining specificity. Overall, our findings suggest a modular organization of knowledge accessed through compositional operations, and advance methods for efficient, targeted LLM manipulation.
- Africa > Nigeria > Federal Capital Territory > Abuja (0.05)
- Asia > China > Beijing > Beijing (0.05)
- South America > Peru (0.04)
- (21 more...)
- Research Report > Experimental Study (1.00)
- Research Report > New Finding (0.68)
- Information Technology > Artificial Intelligence > Representation & Reasoning (1.00)
- Information Technology > Artificial Intelligence > Natural Language > Large Language Model (1.00)
- Information Technology > Artificial Intelligence > Natural Language > Chatbot (0.94)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (0.94)
- North America > Canada (0.14)
- Europe > Croatia (0.14)
- Research Report > Experimental Study (1.00)
- Research Report > New Finding (0.68)
Taming Knowledge Conflicts in Language Models
Li, Gaotang, Chen, Yuzhong, Tong, Hanghang
Language Models (LMs) often encounter knowledge conflicts when parametric memory contradicts contextual knowledge. Previous works attribute this conflict to the interplay between "memory heads" and "context heads", attention heads assumed to promote either memory or context exclusively. In this study, we go beyond this fundamental assumption by uncovering a critical phenomenon we term the "superposition of contextual information and parametric memory", where highly influential attention heads could simultaneously contribute to both memory and context. Building upon this insight, we propose Just Run Twice (JUICE), a test-time attention intervention method that steers LMs toward either parametric beliefs or contextual knowledge without requiring fine-tuning. JUICE identifies a set of reliable attention heads and leverages a dual-run approach to mitigate the superposition effects. Extensive experiments across 11 datasets and 6 model architectures demonstrate that JUICE sets the new state-of-the-art performance and robust generalization, achieving significant and consistent improvement across different domains under various conflict types. Finally, we theoretically analyze knowledge conflict and the superposition of contextual information and parametric memory in attention heads, which further elucidates the effectiveness of JUICE in these settings.
- Asia > China (0.14)
- North America > United States > Illinois (0.14)
- Information Technology > Artificial Intelligence > Representation & Reasoning (1.00)
- Information Technology > Artificial Intelligence > Natural Language > Large Language Model (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (0.92)
- Information Technology > Artificial Intelligence > Natural Language > Chatbot (0.67)
ReAgent: Reversible Multi-Agent Reasoning for Knowledge-Enhanced Multi-Hop QA
Xinjie, Zhao, Gao, Fan, Yang, Rui, Chen, Yingjian, Wang, Yuyang, Zhu, Ying, Tang, Jiacheng, Li, Irene
Recent advances in large language models (LLMs) have significantly improved multi-hop question answering (QA) through direct Chain-of-Thought (CoT) reasoning. However, the irreversible nature of CoT leads to error accumulation, making it challenging to correct mistakes in multi-hop reasoning. This paper introduces ReAgent: a Reversible multi-Agent collaborative framework augmented with explicit backtracking mechanisms, enabling reversible multi-hop reasoning. By incorporating text-based retrieval, information aggregation and validation, our system can detect and correct errors mid-reasoning, leading to more robust and interpretable QA outcomes. The framework and experiments serve as a foundation for future work on error-tolerant QA systems. Empirical evaluations across three benchmarks indicate ReAgent's efficacy, yielding average about 6\% improvements against baseline models.
- North America > United States > California > Los Angeles County > Los Angeles (0.32)
- North America > Canada (0.14)
- Europe > Spain (0.14)
- (3 more...)
- Health & Medicine (0.46)
- Leisure & Entertainment > Sports > Olympic Games (0.31)
Probing Language Models on Their Knowledge Source
Tighidet, Zineddine, Mogini, Andrea, Mei, Jiali, Piwowarski, Benjamin, Gallinari, Patrick
Large Language Models (LLMs) often encounter conflicts between their learned, internal (parametric knowledge, PK) and external knowledge provided during inference (contextual knowledge, CK). Understanding how LLMs models prioritize one knowledge source over the other remains a challenge. In this paper, we propose a novel probing framework to explore the mechanisms governing the selection between PK and CK in LLMs. Using controlled prompts designed to contradict the model's PK, we demonstrate that specific model activations are indicative of the knowledge source employed. We evaluate this framework on various LLMs of different sizes and demonstrate that mid-layer activations, particularly those related to relations in the input, are crucial in predicting knowledge source selection, paving the way for more reliable models capable of handling knowledge conflicts effectively.
- Europe > Croatia (0.14)
- North America > United States > Virginia (0.05)
- Europe > Italy (0.05)
- (17 more...)
Co-occurrence is not Factual Association in Language Models
Pretrained language models can encode a large amount of knowledge and utilize it for various reasoning tasks, yet they can still struggle to learn novel factual knowledge effectively from finetuning on limited textual demonstrations. In this work, we show that the reason for this deficiency is that language models are biased to learn word co-occurrence statistics instead of true factual associations. We identify the differences between two forms of knowledge representation in language models: knowledge in the form of co-occurrence statistics is encoded in the middle layers of the transformer model and does not generalize well to reasoning scenarios beyond simple question answering, while true factual associations are encoded in the lower layers and can be freely utilized in various reasoning tasks. Based on these observations, we propose two strategies to improve the learning of factual associations in language models. We show that training on text with implicit rather than explicit factual associations can force the model to learn factual associations instead of co-occurrence statistics, significantly improving the generalization of newly learned knowledge. We also propose a simple training method to actively forget the learned co-occurrence statistics, which unblocks and enhances the learning of factual associations when training on plain narrative text. On both synthetic and real-world corpora, the two proposed strategies improve the generalization of the knowledge learned during finetuning to reasoning scenarios such as indirect and multi-hop question answering.
- Europe > France (0.05)
- North America > Canada > Ontario > Toronto (0.04)
- Europe > Croatia > Dubrovnik-Neretva County > Dubrovnik (0.04)
Wyoming man running as bot concedes race, launches 'alliance' to inject AI into politics
Wyoming man Victor Miller, who filed mayoral candidacy as AI bot "VIC," speaks out after OpenAI shuts down his account. Victor Miller, who had been running as an artificial intelligence-powered bot named "VIC" [Virtual Integrated Citizen] in Wyoming's capital city, conceded his bid to make technological political history on Wednesday. Miller received 327 votes, or about 3% of the total cast, in Cheyenne's nonpartisan mayoral primary on Tuesday night, according to Laramie County records. On Wednesday, Fox News Digital obtained a statement from Miller saying that he and VIC came up short in their bid to change the definition of political machine in the Cowboy State's capital city: "Today, I, Victor Miller, concede the Cheyenne mayoral race. As the first person to put artificial intelligence directly on the ballot, offering voters the novel choice of AI governance, our campaign has marked a historic moment in politics and technology," Miller said.
- North America > United States > Wyoming > Laramie County > Cheyenne (0.05)
- North America > United States > Pennsylvania (0.05)
- Europe > United Kingdom > England > East Sussex > Brighton (0.05)
Fox News AI Newsletter: US leads world in fastest AI development: report
Fox News chief political anchor Bret Baier has the latest on the pros and cons of the bombshell developments on'Special Report.' TOP OF THE CHARTS: The U.S. topped another study that looked at the fastest-developing artificial intelligence industries in the world, according to a new report. AI ON THE BALLOT: A librarian running as a nonpartisan candidate for mayor of Cheyenne, Wyoming, promises to allow an artificial intelligence bot created by OpenAI to govern the state's capital city. AI POWER PLAY: Google has its eye on the prize -- artificial intelligence -- and it's making a bold power play in the tech arena. The company's recent Made by Google event was more than just showcasing new technology.
- North America > United States > Wyoming > Laramie County > Cheyenne (0.27)
- Asia > China (0.07)
- Media > News (0.93)
- Government (0.77)
- Leisure & Entertainment (0.67)
- Information Technology > Communications > Social Media (1.00)
- Information Technology > Artificial Intelligence > Natural Language > Large Language Model (0.31)
- Information Technology > Artificial Intelligence > Natural Language > Chatbot (0.31)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning > Generative AI (0.31)
How In-Context Learning Emerges from Training on Unstructured Data: On the Role of Co-Occurrence, Positional Information, and Noise Structures
Wibisono, Kevin Christian, Wang, Yixin
Large language models (LLMs) like transformers have impressive in-context learning (ICL) capabilities; they can generate predictions for new queries based on input-output sequences in prompts without parameter updates. While many theories have attempted to explain ICL, they often focus on structured training data similar to ICL tasks, such as regression. In practice, however, these models are trained in an unsupervised manner on unstructured text data, which bears little resemblance to ICL tasks. To this end, we investigate how ICL emerges from unsupervised training on unstructured data. The key observation is that ICL can arise simply by modeling co-occurrence information using classical language models like continuous bag of words (CBOW), which we theoretically prove and empirically validate. Furthermore, we establish the necessity of positional information and noise structure to generalize ICL to unseen data. Finally, we present instances where ICL fails and provide theoretical explanations; they suggest that the ICL ability of LLMs to identify certain tasks can be sensitive to the structure of the training data.
- South America > Suriname > Paramaribo District > Paramaribo (0.04)
- North America > United States > Michigan (0.04)
- Europe > Liechtenstein (0.04)
- (6 more...)
- Information Technology > Artificial Intelligence > Natural Language > Large Language Model (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Statistical Learning (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (1.00)
- (2 more...)