idiom
NLP Datasets for Idiom and Figurative Language Tasks
Matheny, Blake, Nguyen, Phuong Minh, Nguyen, Minh Le, Reynolds, Stephanie
With social media, this informal language has become more easily observable to people and trainers of large language models (LLMs) alike. While the advantage of large corpora seems like the solution to all machine learning and Natural Language Processing (NLP) problems, idioms and figurative language continue to elude LLMs. Finetuning approaches are proving to be optimal, but better and larger datasets can help narrow this gap even further. The datasets presented in this paper provide one answer, while offering a diverse set of categories on which to build new models and develop new approaches. A selection of recent idiom and figurative language datasets were used to acquire a combined idiom list, which was used to retrieve context sequences from a large corpus. One large-scale dataset of potential idiomatic and figurative language expressions and two additional human-annotated datasets of definite idiomatic and figurative language expressions were created to evaluate the baseline ability of pre-trained language models in handling figurative meaning through idiom recognition (detection) tasks. The resulting datasets were post-processed for model agnostic training compatibility, utilized in training, and evaluated on slot labeling and sequence tagging.
- North America > United States > Washington > King County > Seattle (0.14)
- North America > United States > Minnesota > Hennepin County > Minneapolis (0.14)
- Europe > United Kingdom > England > Oxfordshire > Oxford (0.14)
- (11 more...)
- Overview (0.46)
- Research Report (0.41)
A Very Big Fight Over a Very Small Language
In the Swiss Alps, a plan to tidy up Romansh--spoken by less than one per cent of the country--set off a decades-long quarrel over identity, belonging, and the sound of authenticity. After reformers launched Rumantsch Grischun, a standardized version of Romansh's various dialects, traditionalists denounced it as a "bastard," a "castrated" tongue, an act of "linguistic murder." Ask him how it all began, and he remembers the ice. It was a bitter morning in January, 1982, when Bernard Cathomas, aged thirty-six, carefully picked his way up a slippery, sloping Zurich street. His destination was No. 33, an ochre house with green shutters--the home of Heinrich Schmid, a linguist at the University of Zurich. Inside, the décor suggested that "professor" was an encompassing identity: old wooden floors, a faded carpet, a living room seemingly untouched since the nineteen-thirties, when Schmid had grown up in the house. Schmid's wife served, a Swiss carrot cake that manages bourgeois indulgence with a vegetable alibi. Cathomas had already written from Chur, in the canton of the Grisons, having recently become the general secretary of the Lia Rumantscha, a small association charged with protecting Switzerland's least known national language, Romansh. Spoken by less than one per cent of the Swiss population, the language was itself splintered into five major "idioms," not always readily intelligible to one another, each with its own spelling conventions. Earlier attempts at unification had collapsed in rivalries. In his letter, Cathomas said that Schmid's authority would be valuable in standardizing the language. Cathomas wrote in German but started and ended in his native Sursilvan, the biggest of the Romansh idioms: " ." Translation: "I thank you very much for your interest and attention to this problem." Schmid, the man he was counting on, hadn't grown up speaking Romansh; he first learned it in high school, and later worked on the "Dicziunari Rumantsch Grischun," a Romansh dictionary begun in 1904 and still lumbering toward completion.
- Europe > Switzerland > Zürich > Zürich (0.45)
- North America > United States > Texas (0.04)
- North America > United States > New York (0.04)
- (6 more...)
- Media (0.93)
- Government (0.93)
- Education > Educational Setting > K-12 Education > Secondary School (0.48)
Visual Puns from Idioms: An Iterative LLM-T2IM-MLLM Framework
Xiao, Kelaiti, Yang, Liang, Zhang, Dongyu, Tulajiang, Paerhati, Lin, Hongfei
We study idiom-based visual puns--images that align an idiom's literal and figurative meanings--and present an iterative framework that coordinates a large language model (LLM), a text-to-image model (T2IM), and a multimodal LLM (MLLM) for automatic generation and evaluation. Given an idiom, the system iteratively (i) generates detailed visual prompts, (ii) synthesizes an image, (iii) infers the idiom from the image, and (iv) refines the prompt until recognition succeeds or a step limit is reached. Using 1,000 idioms as inputs, we synthesize a corresponding dataset of visual pun images with paired prompts, enabling benchmarking of both generation and understanding. Experiments across 10 LLMs, 10 MLLMs, and one T2IM (Qwen-Image) show that MLLM choice is the primary performance driver: GPT achieves the highest accuracies, Gemini follows, and the best open-source MLLM (Gemma) is competitive with some closed models. On the LLM side, Claude attains the strongest average performance for prompt generation.
Anatomy of an Idiom: Tracing Non-Compositionality in Language Models
We investigate the processing of idiomatic expressions in transformer-based language models using a novel set of techniques for circuit discovery and analysis. First discovering circuits via a modified path patching algorithm, we find that idiom processing exhibits distinct computational patterns. We identify and investigate ``Idiom Heads,'' attention heads that frequently activate across different idioms, as well as enhanced attention between idiom tokens due to earlier processing, which we term ``augmented reception.'' We analyze these phenomena and the general features of the discovered circuits as mechanisms by which transformers balance computational efficiency and robustness. Finally, these findings provide insights into how transformers handle non-compositional language and suggest pathways for understanding the processing of more complex grammatical constructions.
- Europe > Switzerland > Vaud > Lausanne (0.40)
- South America > Colombia > Meta Department > Villavicencio (0.04)
- Europe > Ireland > Leinster > County Dublin > Dublin (0.04)
- Asia > China > Hong Kong (0.04)
- Asia > Middle East > Jordan (0.04)
- North America > Canada (0.04)
- Information Technology > Artificial Intelligence > Natural Language > Grammars & Parsing (1.00)
- Information Technology > Artificial Intelligence > Representation & Reasoning > Uncertainty > Bayesian Inference (0.93)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (0.69)
- Information Technology > Artificial Intelligence > Machine Learning > Learning Graphical Models > Directed Networks > Bayesian Learning (0.68)
Exposing the Cracks: Vulnerabilities of Retrieval-Augmented LLM-based Machine Translation
Sun, Yanming, Zhan, Runzhe, Cheang, Chi Seng, Wu, Han, Liu, Xuebo, Niu, Yuyao, Ye, Fengying, Lan, Kaixin, Chao, Lidia S., Wong, Derek F.
REtrieval-Augmented LLM-based Machine Translation (REAL-MT) shows promise for knowledge-intensive tasks like idiomatic translation, but its reliability under noisy retrieval, a common challenge in real-world deployment, remains poorly understood. To address this gap, we propose a noise synthesis framework and new metrics to systematically evaluate REAL-MT's reliability across high-, medium-, and low-resource language pairs. Using both open-and closed-sourced models, including standard LLMs and large reasoning models (LRMs), we find that models heavily rely on retrieved context, and this dependence is significantly more detrimental in low-resource language pairs, producing nonsensical translations. Although LRMs possess enhanced reasoning capabilities, they show no improvement in error correction and are even more susceptible to noise, tending to rationalize incorrect contexts. Attention analysis reveals a shift from the source idiom to noisy content, while confidence increases despite declining accuracy, indicating poor self-monitoring. To mitigate these issues, we investigate training-free and fine-tuning strategies, which improve robustness at the cost of performance in clean contexts, revealing a fundamental trade-off. Our findings highlight the limitations of current approaches, underscoring the need for self-verifying integration mechanisms.
- Information Technology > Artificial Intelligence > Natural Language > Machine Translation (1.00)
- Information Technology > Artificial Intelligence > Natural Language > Large Language Model (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (0.46)
Beyond Understanding: Evaluating the Pragmatic Gap in LLMs' Cultural Processing of Figurative Language
Attia, Mena, Muhamed, Aashiq, Alkhamissi, Mai, Solorio, Thamar, Diab, Mona
We present a comprehensive evaluation of the ability of large language models (LLMs) to process culturally grounded language, specifically to understand and pragmatically use figurative expressions that encode local knowledge and cultural nuance. Using figurative language as a proxy for cultural nuance and local knowledge, we design evaluation tasks for contextual understanding, pragmatic use, and connotation interpretation in Arabic and English. We evaluate 22 open- and closed-source LLMs on Egyptian Arabic idioms, multidialectal Arabic proverbs, and English proverbs. Our results show a consistent hierarchy: the average accuracy for Arabic proverbs is 4.29% lower than for English proverbs, and performance for Egyptian idioms is 10.28% lower than for Arabic proverbs. For the pragmatic use task, accuracy drops by 14.07% relative to understanding, though providing contextual idiomatic sentences improves accuracy by 10.66%. Models also struggle with connotative meaning, reaching at most 85.58% agreement with human annotators on idioms with 100% inter-annotator agreement. These findings demonstrate that figurative language serves as an effective diagnostic for cultural reasoning: while LLMs can often interpret figurative meaning, they face challenges in using it appropriately. To support future research, we release Kinayat, the first dataset of Egyptian Arabic idioms designed for both figurative understanding and pragmatic use evaluation.
- Europe > Austria > Vienna (0.14)
- Asia > Middle East > UAE > Abu Dhabi Emirate > Abu Dhabi (0.14)
- North America > Canada > Ontario > Toronto (0.04)
- (16 more...)
Visualization Tasks for Unlabelled Graphs
Oddo, Matt I. B., Smith, Ryan, Kobourov, Stephen, Munzner, Tamara
We investigate tasks that can be accomplished with unlabelled graphs, which are graphs with nodes that do not have attached persistent or semantically meaningful labels. New visualization techniques to represent unlabelled graphs have been proposed, but more understanding of unlabelled graph tasks is required before these techniques can be adequately evaluated. Some tasks apply to both labelled and unlabelled graphs, but many do not translate between these contexts. We propose a data abstraction model that distinguishes the Unlabelled context from the increasingly semantically rich Labelled, Attributed, and Augmented contexts. We filter tasks collected and gleaned from the literature according to our data abstraction and analyze the surfaced tasks, leading to a taxonomy of abstract tasks for unlabelled graphs. Our task taxonomy is organized according to the Scope of the data at play, the Action intended by the user, and the Target data under consideration. We show the descriptive power of this task abstraction by connecting to concrete examples from previous frameworks, and connect these abstractions to real-world problems. To showcase the evaluative power of the taxonomy, we perform a preliminary assessment of 6 visualizations for each task. For each combination of task and visual encoding, we consider the effort required from viewers, the likelihood of task success, and how both factors vary between small-scale and large-scale graphs.
- North America > Canada > British Columbia (0.04)
- Europe > Germany > Bavaria > Upper Bavaria > Munich (0.04)
- Overview (0.93)
- Research Report (0.82)
Can generative AI figure out figurative language? The influence of idioms on essay scoring by ChatGPT, Gemini, and Deepseek
The developments in Generative AI technologies have paved the way for numerous innovations in different fields. Recently, Generative AI has been proposed as a competitor to AES systems in evaluating student essays automatically. Considering the potential limitations of AI in processing idioms, this study assessed the scoring performances of Generative AI models for essays with and without idioms by incorporating insights from Corpus Linguistics and Computational Linguistics. Two equal essay lists were created from 348 student essays taken from a corpus: one with multiple idioms present in each essay and another with no idioms in essays. Three Generative AI models (ChatGPT, Gemini, and Deepseek) were asked to score all essays in both lists three times, using the same rubric used by human raters in assigning essay scores. The results revealed excellent consistency for all models, but Gemini outperformed its competitors in interrater reliability with human raters. There was also no detectable bias for any demographic group in AI assessment. For essays with multiple idioms, Gemini followed a the most similar pattern to human raters. While the models in the study demonstrated potential for a hybrid approach, Gemini was the best candidate for the task due to its ability to handle figurative language and showed promise for handling essay-scoring tasks alone in the future.
- North America > United States (0.14)
- Europe > Austria > Vienna (0.14)
- Europe > United Kingdom > England > Oxfordshire > Oxford (0.04)
- Asia > Middle East > Republic of Türkiye (0.04)
- Education > Curriculum > Subject-Specific Education (0.54)
- Education > Assessment & Standards > Student Performance (0.38)