neologism
Neologism Learning for Controllability and Self-Verbalization
Hewitt, John, Tafjord, Oyvind, Geirhos, Robert, Kim, Been
Humans invent new words when there is a rising demand for a new useful concept (e.g., doomscrolling). We explore and validate a similar idea in our communication with LLMs: introducing new words to better understand and control the models, expanding on the recently introduced neologism learning. This method introduces a new word by adding a new word embedding and training with examples that exhibit the concept with no other changes in model parameters. We show that adding a new word allows for control of concepts such as flattery, incorrect answers, text length, as well as more complex concepts in AxBench. We discover that neologisms can also further our understanding of the model via self-verbalization: models can describe what each new word means to them in natural language, like explaining that a word that represents a concept of incorrect answers means ``a lack of complete, coherent, or meaningful answers...'' To validate self-verbalizations, we introduce plug-in evaluation: we insert the verbalization into the context of a model and measure whether it controls the target concept. In some self-verbalizations, we find machine-only synonyms: words that seem unrelated to humans but cause similar behavior in machines. Finally, we show how neologism learning can jointly learn multiple concepts in multiple words.
Hatevolution: What Static Benchmarks Don't Tell Us
Di Bonaventura, Chiara, McGillivray, Barbara, He, Yulan, Meroño-Peñuela, Albert
Language changes over time, including in the hate speech domain, which evolves quickly following social dynamics and cultural shifts. While NLP research has investigated the impact of language evolution on model training and has proposed several solutions for it, its impact on model benchmarking remains under-explored. Yet, hate speech benchmarks play a crucial role to ensure model safety. In this paper, we empirically evaluate the robustness of 20 language models across two evolving hate speech experiments, and we show the temporal misalignment between static and time-sensitive evaluations. Our findings call for time-sensitive linguistic benchmarks in order to correctly and reliably evaluate language models in the hate speech domain.
- Asia > Middle East > UAE > Abu Dhabi Emirate > Abu Dhabi (0.14)
- Asia > Japan > Honshū > Kantō > Tokyo Metropolis Prefecture > Tokyo (0.14)
- North America > Puerto Rico > Peñuelas > Peñuelas (0.05)
- (9 more...)
- Research Report > Experimental Study (0.67)
- Research Report > New Finding (0.48)
NeoN: A Tool for Automated Detection, Linguistic and LLM-Driven Analysis of Neologisms in Polish
Tomaszewska, Aleksandra, Czerski, Dariusz, Żuk, Bartosz, Ogrodniczuk, Maciej
We introduce NeoN, a tool for detecting and analyzing Polish neologisms. Unlike traditional dictionary-based methods requiring extensive manual review, NeoN combines reference corpora, Polish-specific linguistic filters, an LLM-driven precision-boosting filter, and daily RSS monitoring in a multi-layered pipeline. The system uses context-aware lemmatization, frequency analysis, and orthographic normalization to extract candidate neologisms while consolidating inflectional variants. Researchers can verify candidates through an intuitive interface with visualizations and filtering controls. An integrated LLM module automatically generates definitions and categorizes neologisms by domain and sentiment. Evaluations show NeoN maintains high accuracy while significantly reducing manual effort, providing an accessible solution for tracking lexical innovation in Polish.
- Europe > Poland > Masovia Province > Warsaw (0.05)
- Europe > Poland > Greater Poland Province > Poznań (0.04)
Can AI mimic the human ability to define neologisms?
An ongoing and intriguing debate focuses on whether Large Language Models (LLMs) can replicate human language. The literature presents mixed evidence on this matter. Several studies suggest that LLMs can generate text closely resembling human language (Bubeck et al., 2023; Clark et al., 2021; Georgiou, 2025). However, the widely accept ed concept of a universal grammar inherent in humans (Chomsky, 2000) challenges the idea that machine cognition can mirror human cognition. According to Chomsky et al. (2023), models like ChatGPT function as statistical engines driven by pattern recognitio n. Supporting this perspective, other studies highlight significant differences between human cognition and LLMs, which are reflected in language (Cai et al., 2024; Georgiou, 2024; Herbold et al., 2023). For instance, Georgiou (2024) examined how various linguistic components are represented in human - written and AI - generated texts, assessing the ability of ChatGPT to emulate human writing. The author found that d espite AI - generated texts appear ing to mimic human language, the results revealed signifi cant differences across multiple linguistic features in the domains of phonology, grammar, and semantics.
- North America > Canada > Alberta (0.14)
- Europe > Austria > Vienna (0.14)
- Europe > Greece > Central Macedonia > Thessaloniki (0.05)
- (2 more...)
We Can't Understand AI Using our Existing Vocabulary
Hewitt, John, Geirhos, Robert, Kim, Been
This position paper argues that, in order to understand AI, we cannot rely on our existing vocabulary of human words. Instead, we should strive to develop neologisms: new words that represent precise human concepts that we want to teach machines, or machine concepts that we need to learn. We start from the premise that humans and machines have differing concepts. This means interpretability can be framed as a communication problem: humans must be able to reference and control machine concepts, and communicate human concepts to machines. Creating a shared human-machine language through developing neologisms, we believe, could solve this communication problem. Successful neologisms achieve a useful amount of abstraction: not too detailed, so they're reusable in many contexts, and not too high-level, so they convey precise information. As a proof of concept, we demonstrate how a "length neologism" enables controlling LLM response length, while a "diversity neologism" allows sampling more variable responses. Taken together, we argue that we cannot understand AI using our existing vocabulary, and expanding it through neologisms creates opportunities for both controlling and understanding machines better.
- North America > United States > Minnesota > Hennepin County > Minneapolis (0.14)
- North America > United States > Washington > King County > Seattle (0.04)
- North America > United States > Massachusetts > Middlesex County > Cambridge (0.04)
- (3 more...)
Detecting Turkish Synonyms Used in Different Time Periods
Yazar, Umur Togay, Kutlu, Mucahid
Dynamic structure of languages poses significant challenges in applying natural language processing models on historical texts, causing decreased performance in various downstream tasks. Turkish is a prominent example of rapid linguistic transformation due to the language reform in the 20th century. In this paper, we propose two methods for detecting synonyms used in different time periods, focusing on Turkish. In our first method, we use Orthogonal Procrustes method to align the embedding spaces created using documents written in the corresponding time periods. In our second method, we extend the first one by incorporating Spearman's correlation between frequencies of words throughout the years. In our experiments, we show that our proposed methods outperform the baseline method. Furthermore, we observe that the efficacy of our methods remains consistent when the target time period shifts from the 1960s to the 1980s. However, their performance slightly decreases for subsequent time periods.
- Europe > United Kingdom > England > Cambridgeshire > Cambridge (0.04)
- Asia > Middle East > Republic of Türkiye > Ankara Province > Ankara (0.04)
- Asia > Middle East > Qatar > Ad-Dawhah > Doha (0.04)
NEO-BENCH: Evaluating Robustness of Large Language Models with Neologisms
Zheng, Jonathan, Ritter, Alan, Xu, Wei
The performance of Large Language Models (LLMs) degrades from the temporal drift between data used for model training and newer text seen during inference. One understudied avenue of language change causing data drift is the emergence of neologisms -- new word forms -- over time. We create a diverse resource of recent English neologisms by using several popular collection methods. We analyze temporal drift using neologisms by comparing sentences containing new words with near-identical sentences that replace neologisms with existing substitute words. Model performance is nearly halved in machine translation when a single neologism is introduced in a sentence. Motivated by these results, we construct a benchmark to evaluate LLMs' ability to generalize to neologisms with various natural language understanding tasks and model perplexity. Models with later knowledge cutoff dates yield lower perplexities and perform better in downstream tasks. LLMs are also affected differently based on the linguistic origins of words, indicating that neologisms are complex for static LLMs to address. We will release our benchmark and code for reproducing our experiments.
- Asia > Middle East > UAE > Abu Dhabi Emirate > Abu Dhabi (0.04)
- Asia > Singapore (0.04)
- North America > United States > New York > New York County > New York City (0.04)
- (18 more...)
- Information Technology (1.00)
- Health & Medicine > Therapeutic Area > Infections and Infectious Diseases (0.93)
- Media (0.68)
A primer on getting neologisms from foreign languages to under-resourced languages
Neologisms are certain uses, expressions, and words that did not traditionally exist in a language, but are incorporated into it due to the need of speakers to adapt to a new reality [1]. That is, neologisms are those new words and expressions that speakers incorporate into a language, as new things and new ways of doing to name arise. They are the exact opposite of archaisms. The appearance of neologisms is a common and ordinary process in all languages, forced as they are to adapt and update or die. However, a word can be considered a neologism only for a certain time, since once it has been incorporated and normalized as part of the language, it simply ceases to be a novelty. The simplest way to classify neologisms would be from the method used to create them, thus we have: 1. morphological neologisms: they are built using words that already exist in the language, through the processes of composition or derivation. For example, the word "aircraft" was once a neologism, made up of the prefix "air" and the suffix "craft". This also happens with "teleoperators" or with "biosecurity".
- South America > Peru (0.05)
- South America > Bolivia (0.04)
- North America > Mexico (0.04)
- (4 more...)
- Research Report (0.82)
- Overview (0.50)