nominalization
Comparing human and LLM proofreading in L2 writing: Impact on lexical and syntactic features
Sung, Hakyung, Csuros, Karla, Sung, Min-Chang
This study examines the lexical and syntactic interventions of human and LLM proofreading aimed at improving overall intelligibility in identical second language writings, and evaluates the consistency of outcomes across three LLMs (ChatGPT-4o, Llama3.1-8b, Deepseek-r1-8b). Findings show that both human and LLM proofreading enhance bigram lexical features, which may contribute to better coherence and contextual connectedness between adjacent words. However, LLM proofreading exhibits a more generative approach, extensively reworking vocabulary and sentence structures, such as employing more diverse and sophisticated vocabulary and incorporating a greater number of adjective modifiers in noun phrases. The proofreading outcomes are highly consistent in major lexical and syntactic features across the three models.
- North America > United States > Oregon (0.40)
- Asia > Taiwan (0.04)
- Asia > Japan (0.04)
- (13 more...)
Derivational Morphology Reveals Analogical Generalization in Large Language Models
Hofmann, Valentin, Weissweiler, Leonie, Mortensen, David, Schütze, Hinrich, Pierrehumbert, Janet
What mechanisms underlie linguistic generalization in large language models (LLMs)? This question has attracted considerable attention, with most studies analyzing the extent to which the language skills of LLMs resemble rules. As of yet, it is not known whether linguistic generalization in LLMs could equally well be explained as the result of analogical processes, which can be formalized as similarity operations on stored exemplars. A key shortcoming of prior research is its focus on linguistic phenomena with a high degree of regularity, for which rule-based and analogical approaches make the same predictions. Here, we instead examine derivational morphology, specifically English adjective nominalization, which displays notable variability. We introduce a new method for investigating linguistic generalization in LLMs: focusing on GPT-J, we fit cognitive models that instantiate rule-based and analogical learning to the LLM training data and compare their predictions on a set of nonce adjectives with those of the LLM, allowing us to draw direct conclusions regarding underlying mechanisms. As expected, rule-based and analogical models explain the predictions of GPT-J equally well for adjectives with regular nominalization patterns. However, for adjectives with variable nominalization patterns, the analogical model provides a much better match. Furthermore, GPT-J's behavior is sensitive to the individual word frequencies, even for regular forms, a behavior that is consistent with an analogical account of regular forms but not a rule-based one. These findings refute the hypothesis that GPT-J's linguistic generalization on adjective nominalization involves rules, suggesting similarity operations on stored exemplars as the underlying mechanism. Overall, our study suggests that analogical processes play a bigger role in the linguistic generalization of LLMs than previously thought.
- Europe > United Kingdom > England > Oxfordshire > Oxford (0.28)
- Europe > United Kingdom > England > Cambridgeshire > Cambridge (0.14)
- North America > United States > New York > New York County > New York City (0.14)
- (12 more...)
- Research Report > New Finding (1.00)
- Research Report > Experimental Study (1.00)
Paradigm Completion for Derivational Morphology
Cotterell, Ryan, Vylomova, Ekaterina, Khayrallah, Huda, Kirov, Christo, Yarowsky, David
The generation of complex derived word forms has been an overlooked problem in NLP; we fill this gap by applying neural sequence-to-sequence models to the task. We overview the theoretical motivation for a paradigmatic treatment of derivational morphology, and introduce the task of derivational paradigm completion as a parallel to inflectional paradigm completion. State-of-the-art neural models, adapted from the inflection task, are able to learn a range of derivation patterns, and outperform a non-neural baseline by 16.4%. However, due to semantic, historical, and lexical considerations involved in derivational morphology, future work will be needed to achieve performance parity with inflection-generating systems.
- North America > United States > California > Los Angeles County > Los Angeles (0.14)
- Europe > United Kingdom > England > Cambridgeshire > Cambridge (0.05)
- North America > United States > California > San Diego County > San Diego (0.04)
- (12 more...)
The "-ification" of Everything
A half-formed thought feels worse than an empty head--the tip-of-the-tongue sensation, the inkling of a there there without the foggiest notion of how to get, well, there. Especially dire is when the "what" that we wish to articulate feels half-formed itself, something observable yet emergent, for which the masses have yet to find language. But all we have is language, of course, and so we must muddle through, reaching for a word to serve as a placeholder for our idea until something better comes along. Some would say that finding new language is the work of scholars, but in the age of the Internet we may have lost track of who is leading whom. However provisional, the placeholders sometimes stick.
- North America > United States > Illinois > Cook County > Chicago (0.06)
- Oceania > New Zealand (0.05)
- North America > United States > New York (0.05)
- Leisure & Entertainment (1.00)
- Media > Television (0.30)
- Health & Medicine > Therapeutic Area > Infections and Infectious Diseases (0.30)
- Health & Medicine > Therapeutic Area > Immunology (0.30)
Unsupervised Mapping of Arguments of Deverbal Nouns to Their Corresponding Verbal Labels
Weinstein, Aviv, Goldberg, Yoav
Deverbal nouns are nominal forms of verbs commonly used in written English texts to describe events or actions, as well as their arguments. However, many NLP systems, and in particular pattern-based ones, neglect to handle such nominalized constructions. The solutions that do exist for handling arguments of nominalized constructions are based on semantic annotation and require semantic ontologies, making their applications restricted to a small set of nouns. We propose to adopt instead a more syntactic approach, which maps the arguments of deverbal nouns to the universal-dependency relations of the corresponding verbal construction. We present an unsupervised mechanism -- based on contextualized word representations -- which allows to enrich universal-dependency trees with dependency arcs denoting arguments of deverbal nouns, using the same labels as the corresponding verbal cases. By sharing the same label set as in the verbal case, patterns that were developed for verbs can be applied without modification but with high accuracy also to the nominal constructions.
- North America > United States > Pennsylvania (0.04)
- North America > Dominican Republic (0.04)
- Asia > Singapore (0.04)
QASem Parsing: Text-to-text Modeling of QA-based Semantics
Klein, Ayal, Hirsch, Eran, Eliav, Ron, Pyatkin, Valentina, Caciularu, Avi, Dagan, Ido
Several recent works have suggested to represent semantic relations with questions and answers, decomposing textual information into separate interrogative natural language statements. In this paper, we consider three QA-based semantic tasks - namely, QA-SRL, QANom and QADiscourse, each targeting a certain type of predication - and propose to regard them as jointly providing a comprehensive representation of textual information. To promote this goal, we investigate how to best utilize the power of sequence-to-sequence (seq2seq) pre-trained language models, within the unique setup of semi-structured outputs, consisting of an unordered set of question-answer pairs. We examine different input and output linearization strategies, and assess the effect of multitask learning and of simple data augmentation techniques in the setting of imbalanced training data. Consequently, we release the first unified QASem parsing tool, practical for downstream applications who can benefit from an explicit, QA-based account of information units in a text.
- North America > United States > Minnesota > Hennepin County > Minneapolis (0.14)
- Europe > United Kingdom > Scotland (0.04)
- North America > United States > Louisiana > Orleans Parish > New Orleans (0.04)
- (11 more...)
Last Words: Computational Linguistics and Deep Learning - MITP on Nautilus
Deep Learning waves have lapped at the shores of computational linguistics for several years now, but 2015 seems like the year when the full force of the tsunami hit the major Natural Language Processing (NLP) conferences. However, some pundits are predicting that the final damage will be even worse. Accompanying ICML 2015 in Lille, France, there was another, almost as big, event: the 2015 Deep Learning Workshop. The workshop ended with a panel discussion, and at it, Neil Lawrence said, "NLP is kind of like a rabbit in the headlights of the Deep Learning machine, waiting to be flattened." Now that is a remark that the computational linguistics community has to take seriously!
- Europe > France > Hauts-de-France > Nord > Lille (0.24)
- North America > United States > Massachusetts > Middlesex County > Cambridge (0.05)
- Asia > Middle East > Jordan (0.05)
- (7 more...)