Goto

Collaborating Authors

 Radev, Dragomir R.


What Should I Learn First: Introducing LectureBank for NLP Education and Prerequisite Chain Learning

arXiv.org Machine Learning

Recent years have witnessed the rising popularity of Natural Language Processing (NLP) and related fields such as Artificial Intelligence (AI) and Machine Learning (ML). Many online courses and resources are available even for those without a strong background in the field. Often the student is curious about a specific topic but does not quite know where to begin studying. To answer the question of "what should one learn first," we apply an embedding-based method to learn prerequisite relations for course concepts in the domain of NLP. We introduce LectureBank, a dataset containing 1,352 English lecture files collected from university courses which are each classified according to an existing taxonomy as well as 208 manually-labeled prerequisite relation topics, which is publicly available. The dataset will be useful for educational purposes such as lecture preparation and organization as well as applications such as reading list generation. Additionally, we experiment with neural graph-based networks and non-neural classifiers to learn these prerequisite relations from our dataset.


Zero-shot Transfer Learning for Semantic Parsing

arXiv.org Machine Learning

While neural networks have shown impressive performance on large datasets, applying these models to tasks where little data is available remains a challenging problem. In this paper we propose to use feature transfer in a zero-shot experimental setting on the task of semantic parsing. We first introduce a new method for learning the shared space between multiple domains based on the prediction of the domain label for each example. Our experiments support the superiority of this method in a zero-shot experimental setting in terms of accuracy metrics compared to state-of-the-art techniques. In the second part of this paper we study the impact of individual domains and examples on semantic parsing performance. We use influence functions to this aim and investigate the sensitivity of domain-label classification loss on each example. Our findings reveal that cross-domain adversarial attacks identify useful examples for training even from the domains the least similar to the target domain. Augmenting our training data with these influential examples further boosts our accuracy at both the token and the sequence level.


Exploiting Phase Transition in Latent Networks for Clustering

AAAI Conferences

In this paper, we model the pair-wise similarities of a setof documents as a weighted network with a single cutoffparameter. Such a network can be thought of an ensemble of unweighted graphs, each consisting of edges withweights greater than the cutoff value. We look at this network ensemble as a complex system with a temperature parameter, and refer to it as a Latent Network. Ourexperiments on a number of datasets from two different domains show that certain properties of latent networks like clustering coefficient, average shortest path,and connected components exhibit patterns that are significantly divergent from randomized networks. We explain that these patterns reflect the network phase transition as well as the existence of a community structure in document collections. Using numerical analysis,we show that we can use the aforementioned networkproperties to predicts the clustering Normalized MutualInformation (NMI) with high correlation (rho > 0.9). Finally we show that our clustering method significantlyoutperforms other baseline methods (NMI > 0.5)


Networks and Natural Language Processing

AI Magazine

Over the last few years, a number of areas of natural language processing have begun applying graph-based techniques. These include, among others, text summarization, syntactic parsing, word-sense disambiguation, ontology construction, sentiment and subjectivity analysis, and text clustering. In this paper, we present some of the most successful graph-based representations and algorithms used in language processing and try to explain how and why they work.


Introduction to the Special Issue on AI and Networks

AI Magazine

This introduction to AI Magazine's Special Issueon Networks and AI summarizes the seven articles in thespecial issue by characterizing the nature of thenetworks that are the focus of each of the papers.A short tutorial on graph theory and network structuresis included for those less familiar with the topic.


Networks and Natural Language Processing

AI Magazine

Over the last few years, a number of areas of natural language processing have begun applying graph-based techniques. These include, among others, text summarization, syntactic parsing, word-sense disambiguation, ontology construction, sentiment and subjectivity analysis, and text clustering. In this paper, we present some of the most successful graph-based representations and algorithms used in language processing and try to explain how and why they work.


The AAAI Spring Symposia

AI Magazine

The Association for the Advancement of Artificial Intelligence, in cooperation with Stanford University's Department of Computer Science, held the 1998 Spring Symposium Series on 23 to 25 March at Stanford University. The topics of the eight symposia were (1) Applying Machine Learning to Discourse Processing, (2) Integrating Robotic Research: Taking the Next Leap, (3) Intelligent Environments, (4) Intelligent Text Summarization, (5) Interactive and Mixed-Initiative Decision-Theoretic Systems, (6) Multimodal Reasoning, (7) Prospects for a Common-Sense Theory of Causation, and (8) Satisficing Models.


The AAAI Spring Symposia

AI Magazine

The Association for the Advancement of Artificial Intelligence, in cooperation with Stanford University's Department of Computer Science, held the 1998 Spring Symposium Series on 23 to 25 March at Stanford University. The topics of the eight symposia were (1) Applying Machine Learning to Discourse Processing, (2) Integrating Robotic Research: Taking the Next Leap, (3) Intelligent Environments, (4) Intelligent Text Summarization, (5) Interactive and Mixed-Initiative Decision-Theoretic Systems, (6) Multimodal Reasoning, (7) Prospects for a Common-Sense Theory of Causation, and (8) Satisficing Models.