Goto

Collaborating Authors

 sahlgren


Large Language Models Preserve Semantic Isotopies in Story Continuations

Cavazza, Marc

arXiv.org Artificial Intelligence

In this work, we explore the relevance of textual semantics to Large Language Models (LLMs), extending previous insights into the connection between distributional semantics and structural semantics. We investigate whether LLM-generated texts preserve semantic isotopies. We design a story continuation experiment using 10,000 ROCStories prompts completed by five LLMs. We first validate GPT-4o's ability to extract isotopies from a linguistic benchmark, then apply it to the generated stories. We then analyze structural (coverage, density, spread) and semantic properties of isotopies to assess how they are affected by completion. Results show that LLM completion within a given token horizon preserves semantic isotopies across multiple properties.


The King's Swedish: AI Rewrites the Book in Scandinavia

#artificialintelligence

If the King of Sweden wants help drafting his annual Christmas speech this year, he could ask the same AI model that's available to his 10 million subjects. As a test, researchers prompted the model, called GPT-SW3, to draft one of the royal messages, and it did a pretty good job, according to Magnus Sahlgren, who heads research in natural language understanding at AI Sweden, a consortium kickstarting the country's journey into the machine learning era. "Later, our minister of digitalization visited us and asked the model to generate arguments for political positions and it came up with some really clever ones -- and he intuitively understood how to prompt the model to generate good text," Sahlgren said. Early successes inspired work on an even larger and more powerful version of the language model they hope will serve any citizen, company or government agency in Scandinavia. The current version packs 3.6 billion parameters and is smart enough to do a few cool things in Swedish.


Semantic Oscillations: Encoding Context and Structure in Complex Valued Holographic Vectors

Vine, Lance De (Queensland University of Technology) | Bruza, Peter (Queensland University of Technology)

AAAI Conferences

In computational linguistics, information retrieval and applied cognition, words and concepts are often represented as vectors in high dimensional spaces computed from a corpus of text. These high dimensional spaces are often referred to as Semantic Spaces. We describe a novel and efficient approach to computing these semantic spaces via the use of complex valued vector representations. We report on the practical implementation of the proposed method and some associated experiments. We also briefly discuss how the proposed system relates to previous theoretical work in Information Retrieval and Quantum Mechanics and how the notions of probability, logic and geometry are integrated within a single Hilbert space representation. In this sense the proposed system has more general application and gives rise to a variety of opportunities for future research.