coherence model
- Asia > South Korea > Seoul > Seoul (0.04)
- Europe > Czechia > South Moravian Region > Brno (0.04)
- Information Technology > Sensing and Signal Processing > Image Processing (1.00)
- Information Technology > Artificial Intelligence > Representation & Reasoning (1.00)
- Information Technology > Artificial Intelligence > Natural Language (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (0.71)
Expressing an Image Stream with a Sequence of Natural Sentences
We propose an approach for retrieving a sequence of natural sentences for an image stream. Since general users often take a series of pictures on their special moments, it would better take into consideration of the whole image stream to produce natural language descriptions. While almost all previous studies have dealt with the relation between a single image and a single natural sentence, our work extends both input and output dimension to a sequence of images and a sequence of sentences. To this end, we design a multimodal architecture called coherent recurrent convolutional network (CRCN), which consists of convolutional neural networks, bidirectional recurrent neural networks, and an entity-based local coherence model. Our approach directly learns from vast user-generated resource of blog posts as text-image parallel training data. We demonstrate that our approach outperforms other state-of-the-art candidate methods, using both quantitative measures (e.g.
- North America > United States > New York (0.05)
- Asia > South Korea > Seoul > Seoul (0.04)
- Europe > Czechia > South Moravian Region > Brno (0.04)
A Digital Language Coherence Marker for Monitoring Dementia
Gkoumas, Dimitris, Tsakalidis, Adam, Liakata, Maria
The use of spontaneous language to derive appropriate digital markers has become an emergent, promising and non-intrusive method to diagnose and monitor dementia. Here we propose methods to capture language coherence as a cost-effective, human-interpretable digital marker for monitoring cognitive changes in people with dementia. We introduce a novel task to learn the temporal logical consistency of utterances in short transcribed narratives and investigate a range of neural approaches. We compare such language coherence patterns between people with dementia and healthy controls and conduct a longitudinal evaluation against three clinical bio-markers to investigate the reliability of our proposed digital coherence marker. The coherence marker shows a significant difference between people with mild cognitive impairment, those with Alzheimer's Disease and healthy controls. Moreover our analysis shows high association between the coherence marker and the clinical bio-markers as well as generalisability potential to other related conditions.
- Europe > United Kingdom > England > Greater London > London (0.04)
- North America > United States > Virginia > Arlington County > Arlington (0.04)
- North America > United States > Pennsylvania > Philadelphia County > Philadelphia (0.04)
- Europe > United Kingdom > England > Cambridgeshire > Cambridge (0.04)
- Research Report > Experimental Study (1.00)
- Research Report > New Finding (0.68)
- Health & Medicine > Therapeutic Area > Neurology > Dementia (1.00)
- Health & Medicine > Therapeutic Area > Neurology > Alzheimer's Disease (1.00)
Understanding Topic Coherence Measures
Topic Modeling is one of the most important NLP fields. It aims to explain a textual dataset by decomposing it into two distributions: topics and words. So, a Topic Modeling Algorithm is a mathematical/statistical model used to infer what are the topics that better represent the data. For simplicity, a topic can be described as a collection of words, like ['ball', 'cat', 'house'] and ['airplane', 'clouds'], but in practice, what an algorithm does is assign each word in our vocabulary a'participation' value in a given topic. The words with the highest values can be considered as the true participants of a topic.
Analyzing Neural Discourse Coherence Models
Farag, Youmna, Valvoda, Josef, Yannakoudakis, Helen, Briscoe, Ted
Different theories have been proposed model's ability to rank a well-organized document to describe the properties that contribute to higher than its noisy counterparts created by discourse coherence and some have been integrated corrupting sentence order in the original document with computational models for empirical (binary discrimination task), and neural evaluation. A popular approach is the entitybased models have achieved remarkable accuracy on model which hypothesizes that coherence this task. Recent efforts have targeted additional can be assessed in terms of the distribution of tasks such as recovering the correct sentence and transitions between entities in a text - by order (Logeswaran et al., 2018; Cui et al., 2018), constructing an entity-grid (Egrid) representation evaluating on realistic data (Lai and Tetreault, (Barzilay and Lapata, 2005, 2008), building 2018; Farag and Yannakoudakis, 2019) and on Centering Theory (Grosz et al., 1995). Subsequent focusing on open-domain models of coherence work has adapted and further extended (Li and Jurafsky, 2017; Xu et al., 2019). Egrid representations (Filippova and Strube, However, less attention has been directed to 2007; Burstein et al., 2010; Elsner and Charniak, investigating and analyzing the properties of coherence 2011; Guinaudeau and Strube, 2013). Other that current models can capture, nor what research has focused on syntactic patterns knowledge is encoded in their representations and that cooccur in text (Louis and Nenkova, how it might relate to aspects of coherence.
- Europe > United Kingdom > England > Cambridgeshire > Cambridge (0.14)
- North America > United States > Minnesota > Hennepin County > Minneapolis (0.14)
- North America > United States > California > San Francisco County > San Francisco (0.14)
- (13 more...)
An Entity-Driven Framework for Abstractive Summarization
Sharma, Eva, Huang, Luyang, Hu, Zhe, Wang, Lu
Abstractive summarization systems aim to produce more coherent and concise summaries than their extractive counterparts. Popular neural models have achieved impressive results for single-document summarization, yet their outputs are often incoherent and unfaithful to the input. In this paper, we introduce SENECA, a novel System for ENtity-drivEn Coherent Abstractive summarization framework that leverages entity information to generate informative and coherent abstracts. Our framework takes a two-step approach: (1) an entity-aware content selection module first identifies salient sentences from the input, then (2) an abstract generation module conducts cross-sentence information compression and abstraction to generate the final summary, which is trained with rewards to promote coherence, conciseness, and clarity. The two components are further connected using reinforcement learning. Automatic evaluation shows that our model significantly outperforms previous state-of-the-art on ROUGE and our proposed coherence measures on New York Times and CNN/Daily Mail datasets. Human judges further rate our system summaries as more informative and coherent than those by popular summarization models.
- Europe > Jersey (0.14)
- North America > United States > Massachusetts > Suffolk County > Boston (0.14)
- Europe > Ireland (0.14)
- (15 more...)
- Media > Music (1.00)
- Leisure & Entertainment (1.00)
- Law (1.00)
- (3 more...)
A Unified Neural Coherence Model
Moon, Han Cheol, Mohiuddin, Tasnim, Joty, Shafiq, Chi, Xu
Recently, neural approaches to coherence modeling have achieved state-of-the-art results in several evaluation tasks. However, we show that most of these models often fail on harder tasks with more realistic application scenarios. In particular, the existing models underperform on tasks that require the model to be sensitive to local contexts such as candidate ranking in conversational dialogue and in machine translation. In this paper, we propose a unified coherence model that incorporates sentence grammar, inter-sentence coherence relations, and global coherence patterns into a common neural framework. With extensive experiments on local and global discrimination tasks, we demonstrate that our proposed model outperforms existing models by a good margin, and establish a new state-of-the-art. 1 Introduction Coherence modeling involves building text analysis models that can distinguish a coherent text from incoherent ones. It has been a key problem in discourse analysis with applications in text generation, summarization, and coherence scoring. V arious linguistic theories have been proposed to formulate coherence, some of which have inspired development of many of the existing coherence models. These include the entity-based local models (Barzilay and Lapata, 2008; Elsner and Charniak, 2011b) that consider syntactic realization of entities in adjacent sentences, inspired by the Centering Theory (Grosz et al., 1995). Another line of research uses discourse relations between sentences to predict local coherence (Pitler and Nenkova, 2008; Lin et al., 2011). These methods are inspired by the discourse structure theories like Rhetorical Structure Theory (RST) (Mann and Thompson, 1988) that formalizes coherence in *Equal contribution terms of discourse relations.
- North America > United States > California > San Francisco County > San Francisco (0.14)
- Asia > Singapore (0.04)
- North America > United States > Oregon > Multnomah County > Portland (0.04)
- (14 more...)
Learning to Extract Coherent Summary via Deep Reinforcement Learning
Wu, Yuxiang (Hong Kong University of Science and Technology) | Hu, Baotian (University of Massachusetts Medical School)
Coherence plays a critical role in producing a high-quality summary from a document. In recent years, neural extractive summarization is becoming increasingly attractive. However, most of them ignore the coherence of summaries when extracting sentences. As an effort towards extracting coherent summaries, we propose a neural coherence model to capture the cross-sentence semantic and syntactic coherence patterns. The proposed neural coherence model obviates the need for feature engineering and can be trained in an end-to-end fashion using unlabeled data. Empirical results show that the proposed neural coherence model can efficiently capture the cross-sentence coherence patterns. Using the combined output of the neural coherence model and ROUGE package as the reward, we design a reinforcement learning method to train a proposed neural extractive summarizer which is named Reinforced Neural Extractive Summarization (RNES) model. The RNES model learns to optimize coherence and informative importance of the summary simultaneously. The experimental results show that the proposed RNES outperforms existing baselines and achieves state-of-the-art performance in term of ROUGE on CNN/Daily Mail dataset. The qualitative evaluation indicates that summaries produced by RNES are more coherent and readable.
- Asia > Nepal (0.04)
- Asia > China > Hong Kong (0.04)
- North America > United States > Massachusetts > Middlesex County > Cambridge (0.04)
- (2 more...)
Expressing an Image Stream with a Sequence of Natural Sentences
We propose an approach for generating a sequence of natural sentences for an image stream. Since general users usually take a series of pictures on their special moments, much online visual information exists in the form of image streams, for which it would better take into consideration of the whole set to generate natural language descriptions. While almost all previous studies have dealt with the relation between a single image and a single natural sentence, our work extends both input and output dimension to a sequence of images and a sequence of sentences. To this end, we design a novel architecture called coherent recurrent convolutional network (CRCN), which consists of convolutional networks, bidirectional recurrent networks, and entity-based local coherence model. Our approach directly learns from vast user-generated resource of blog posts as text-image parallel training data. We demonstrate that our approach outperforms other state-of-the-art candidate methods, using both quantitative measures (e.g. BLEU and top-K recall) and user studies via Amazon Mechanical Turk.
- North America > United States > New York (0.05)
- Asia > South Korea > Seoul > Seoul (0.04)
- Europe > Czechia > South Moravian Region > Brno (0.04)
Generating Chinese Classical Poems with Statistical Machine Translation Models
He, Jing (Tsinghua University) | Zhou, Ming (Microsoft Research Asia) | Jiang, Long (Microsoft Research Asia)
This paper describes a statistical approach to generation of Chinese classical poetry and proposes a novel method to automatically evaluate poems. The system accepts a set of keywords representing the writing intents from a writer and generates sentences one by one to form a completed poem. A statistical machine translation (SMT) system is applied to generate new sentences, given the sentences generated previously. For each line of sentence a specific model specially trained for that line is used, as opposed to using a single model for all sentences. To enhance the coherence of sentences on every line, a coherence model using mutual information is applied to select candidates with better consistency with previous sentences. In addition, we demonstrate the effectiveness of the BLEU metric for evaluation with a novel method of generating diverse references.
- Asia > China > Beijing > Beijing (0.05)
- North America > United States > Pennsylvania > Philadelphia County > Philadelphia (0.04)
- North America > United States > Massachusetts > Middlesex County > Reading (0.04)
- (3 more...)