novel word
The mutual exclusivity bias of bilingual visually grounded speech models
Oneata, Dan, Nortje, Leanne, Matusevych, Yevgen, Kamper, Herman
Mutual exclusivity (ME) is a strategy where a novel word is associated with a novel object rather than a familiar one, facilitating language learning in children. Recent work has found an ME bias in a visually grounded speech (VGS) model trained on English speech with paired images. But ME has also been studied in bilingual children, who may employ it less due to cross-lingual ambiguity. We explore this pattern computationally using bilingual VGS models trained on combinations of English, French, and Dutch. We find that bilingual models generally exhibit a weaker ME bias than monolingual models, though exceptions exist. Analyses show that the combined visual embeddings of bilingual models have a smaller variance for familiar data, partly explaining the increase in confusion between novel and familiar concepts. We also provide new insights into why the ME bias exists in VGS models in the first place. Code and data: https://github.com/danoneata/me-vgs
- North America > United States (0.14)
- North America > Canada > Quebec > Montreal (0.04)
- Europe > Romania > București - Ilfov Development Region > Municipality of Bucharest > Bucharest (0.04)
- (2 more...)
Modelling Child Learning and Parsing of Long-range Syntactic Dependencies
Mahon, Louis, Johnson, Mark, Steedman, Mark
This work develops a probabilistic child language acquisition model to learn a range of linguistic phenonmena, most notably long-range syntactic dependencies of the sort found in object wh-questions, among other constructions. The model is trained on a corpus of real child-directed speech, where each utterance is paired with a logical form as a meaning representation. It then learns both word meanings and language-specific syntax simultaneously. After training, the model can deduce the correct parse tree and word meanings for a given utterance-meaning pair, and can infer the meaning if given only the utterance. The successful modelling of long-range dependencies is theoretically important because it exploits aspects of the model that are, in general, trans-context-free.
- North America > United States > Massachusetts > Middlesex County > Cambridge (0.04)
- North America > United States > New York (0.04)
- Oceania > Australia (0.04)
- (9 more...)
- Information Technology > Artificial Intelligence > Natural Language > Grammars & Parsing (1.00)
- Information Technology > Artificial Intelligence > Representation & Reasoning > Uncertainty > Bayesian Inference (0.67)
- Information Technology > Artificial Intelligence > Machine Learning > Learning Graphical Models > Directed Networks > Bayesian Learning (0.67)
Visually Grounded Speech Models have a Mutual Exclusivity Bias
Nortje, Leanne, Oneaţă, Dan, Matusevych, Yevgen, Kamper, Herman
When children learn new words, they employ constraints such as the mutual exclusivity (ME) bias: a novel word is mapped to a novel object rather than a familiar one. This bias has been studied computationally, but only in models that use discrete word representations as input, ignoring the high variability of spoken words. We investigate the ME bias in the context of visually grounded speech models that learn from natural images and continuous speech audio. Concretely, we train a model on familiar words and test its ME bias by asking it to select between a novel and a familiar object when queried with a novel word. To simulate prior acoustic and visual knowledge, we experiment with several initialisation strategies using pretrained speech and vision networks. Our findings reveal the ME bias across the different initialisation approaches, with a stronger bias in models with more prior (in particular, visual) knowledge. Additional tests confirm the robustness of our results, even when different loss functions are considered.
- Europe > Romania > București - Ilfov Development Region > Municipality of Bucharest > Bucharest (0.04)
- Europe > Netherlands (0.04)
- Africa > South Africa (0.04)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks (1.00)
- Information Technology > Artificial Intelligence > Speech (0.93)
- Information Technology > Artificial Intelligence > Vision (0.93)
- Information Technology > Artificial Intelligence > Representation & Reasoning (0.66)
MEWL: Few-shot multimodal word learning with referential uncertainty
Jiang, Guangyuan, Xu, Manjie, Xin, Shiji, Liang, Wei, Peng, Yujia, Zhang, Chi, Zhu, Yixin
Without explicit feedback, humans can rapidly learn the meaning of words. Children can acquire a new word after just a few passive exposures, a process known as fast mapping. This word learning capability is believed to be the most fundamental building block of multimodal understanding and reasoning. Despite recent advancements in multimodal learning, a systematic and rigorous evaluation is still missing for human-like word learning in machines. To fill in this gap, we introduce the MachinE Word Learning (MEWL) benchmark to assess how machines learn word meaning in grounded visual scenes. MEWL covers human's core cognitive toolkits in word learning: cross-situational reasoning, bootstrapping, and pragmatic learning. Specifically, MEWL is a few-shot benchmark suite consisting of nine tasks for probing various word learning capabilities. These tasks are carefully designed to be aligned with the children's core abilities in word learning and echo the theories in the developmental literature. By evaluating multimodal and unimodal agents' performance with a comparative analysis of human performance, we notice a sharp divergence in human and machine word learning. We further discuss these differences between humans and machines and call for human-like few-shot word learning in machines.
- Asia > China > Beijing > Beijing (0.04)
- North America > United States > Hawaii > Honolulu County > Honolulu (0.04)
- Research Report > Experimental Study (1.00)
- Research Report > New Finding (0.92)
- Education (1.00)
- Health & Medicine > Therapeutic Area (0.67)
- Information Technology > Artificial Intelligence > Representation & Reasoning (1.00)
- Information Technology > Artificial Intelligence > Natural Language > Large Language Model (1.00)
- Information Technology > Artificial Intelligence > Cognitive Science (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (0.93)
A Bayesian Framework for Cross-Situational Word-Learning
For infants, early word learning is a chicken-and-egg problem. One way to learn a word is to observe that it co-occurs with a particular referent across different situations. Another way is to use the social context of an utterance to infer the in- tended referent of a word. Here we present a Bayesian model of cross-situational word learning, and an extension of this model that also learns which social cues are relevant to determining reference. We test our model on a small corpus of mother-infant interaction and find it performs better than competing models. Fi- nally, we show that our model accounts for experimental phenomena including mutual exclusivity, fast-mapping, and generalization from social cues.
Introduction of a novel word embedding approach based on technology labels extracted from patent data
Standke, Mark, Kiwan, Abdullah, Lange, Annalena, Berg, Silvan
Diversity in patent language is growing and makes finding synonyms for conducting patent searches more and more challenging. In addition to that, most approaches for dealing with diverse patent language are based on manual search and human intuition. In this paper, a word embedding approach using statistical analysis of human labeled data to produce accurate and language independent word vectors for technical terms is introduced. This paper focuses on the explanation of the idea behind the statistical analysis and shows first qualitative results. The resulting algorithm is a development of the former EQMania UG (eqmania.com)
- North America > United States (0.04)
- Europe > Switzerland (0.04)
- Europe > Germany (0.04)
- (2 more...)
an overview of GPT-3: AI of the future
OpenAI's paper describes many incredible tasks that GPT-3 can accomplish. For instance, given some text input, it can predict what should come next alarmingly well. In one study, test subjects were asked to differentiate between 500 word long articles written by humans or articles written by GTP-3. When GPT-3 is used with 175 billion parameters, they were only able to correctly identify text from other humans at a rate of 52%. That implies GPT-3 at maximum capacity can nearly replicate human written articles!
- Information Technology > Artificial Intelligence > Natural Language > Large Language Model (1.00)
- Information Technology > Artificial Intelligence > Natural Language > Chatbot (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning > Generative AI (0.30)
Learning Semantic Representations for Novel Words: Leveraging Both Form and Context
Schick, Timo, Schütze, Hinrich
Word embeddings are a key component of high-performing natural language processing (NLP) systems, but it remains a challenge to learn good representations for novel words on the fly, i.e., for words that did not occur in the training data. The general problem setting is that word embeddings are induced on an unlabeled training corpus and then a model is trained that embeds novel words into this induced embedding space. Currently, two approaches for learning embeddings of novel words exist: (i) learning an embedding from the novel word's surface-form (e.g., subword n-grams) and (ii) learning an embedding from the context in which it occurs. In this paper, we propose an architecture that leverages both sources of information - surface-form and context - and show that it results in large increases in embedding quality. Our architecture obtains state-of-the-art results on the Definitional Nonce and Contextual Rare Words datasets. As input, we only require an embedding set and an unlabeled corpus for training our architecture to produce embeddings appropriate for the induced embedding space. Thus, our model can easily be integrated into any existing NLP system and enhance its capability to handle novel words.
- South America > Colombia > Meta Department > Villavicencio (0.05)
- Europe > Germany > Bavaria > Upper Bavaria > Munich (0.04)
Squirrel! Dog MRI study reveals canines really CAN understand some of what their owners say
While many dogs owners think their pets really do understand them, a new study has found they really do have a'rudimentary' understands of words. It could help explain the'squirrel phenomenon' where dogs instantly perk up, become agigated and even start hunting for squirrels when their owner tells them one is close by. However, researchers have been unclear what is actually happening in the canine brain - and how much they really understand. Eddie, a golden retriever-Labrador mix, was part of the study, along with his toys Piggy and Monkey. For the study, 12 dogs of varying breeds were trained for months by their owners to retrieve two different objects, based on the objects' names.
- North America > Canada > Newfoundland and Labrador > Labrador (0.26)
- Europe > Hungary > Budapest > Budapest (0.05)
Necessary and Sufficient Conditions and a Provably Efficient Algorithm for Separable Topic Discovery
Ding, Weicong, Ishwar, Prakash, Saligrama, Venkatesh
We develop necessary and sufficient conditions and a novel provably consistent and efficient algorithm for discovering topics (latent factors) from observations (documents) that are realized from a probabilistic mixture of shared latent factors that have certain properties. Our focus is on the class of topic models in which each shared latent factor contains a novel word that is unique to that factor, a property that has come to be known as separability. Our algorithm is based on the key insight that the novel words correspond to the extreme points of the convex hull formed by the row-vectors of a suitably normalized word co-occurrence matrix. We leverage this geometric insight to establish polynomial computation and sample complexity bounds based on a few isotropic random projections of the rows of the normalized word co-occurrence matrix. Our proposed random-projections-based algorithm is naturally amenable to an efficient distributed implementation and is attractive for modern web-scale distributed data mining applications.
- North America > United States > Georgia > Fulton County > Atlanta (0.04)
- South America > Paraguay > Asunción > Asunción (0.04)
- Asia > Middle East > Jordan (0.04)
- (5 more...)