Xu, Jun
Locally Smoothed Neural Networks
Pang, Liang, Lan, Yanyan, Xu, Jun, Guo, Jiafeng, Cheng, Xueqi
Convolutional Neural Networks (CNN) and the locally connected layer are limited in capturing the importance and relations of different local receptive fields, which are often crucial for tasks such as face verification, visual question answering, and word sequence prediction. To tackle the issue, we propose a novel locally smoothed neural network (LSNN) in this paper. The main idea is to represent the weight matrix of the locally connected layer as the product of the kernel and the smoother, where the kernel is shared over different local receptive fields, and the smoother is for determining the importance and relations of different local receptive fields. Specifically, a multi-variate Gaussian function is utilized to generate the smoother, for modeling the location relations among different local receptive fields. Furthermore, the content information can also be leveraged by setting the mean and precision of the Gaussian function according to the content. Experiments on some variant of MNIST clearly show our advantages over CNN and locally connected layer.
Inside Out: Two Jointly Predictive Models for Word Representations and Phrase Representations
Sun, Fei (Institute of Computing Technology, Chinese Academy of Sciences) | Guo, Jiafeng (Institute of Computing Technology, Chinese Academy of Sciences) | Lan, Yanyan (Institute of Computing Technology, Chinese Academy of Sciences) | Xu, Jun (Institute of Computing Technology, Chinese Academy of Sciences) | Cheng, Xueqi (Institute of Computing Technology, Chinese Academy of Sciences)
Distributional hypothesis lies in the root of most existing word representation models by inferring word meaning from its external contexts. However, distributional models cannot handle rare and morphologically complex words very well and fail to identify some fine-grained linguistic regularity as they are ignoring the word forms. On the contrary, morphology points out that words are built from some basic units, i.e., morphemes. Therefore, the meaning and function of such rare words can be inferred from the words sharing the same morphemes, and many syntactic relations can be directly identified based on the word forms. However, the limitation of morphology is that it cannot infer the relationship between two words that do not share any morphemes. Considering the advantages and limitations of both approaches, we propose two novel models to build better word representations by modeling both external contexts and internal morphemes in a jointly predictive way, called BEING and SEING. These two models can also be extended to learn phrase representations according to the distributed morphology theory. We evaluate the proposed models on similarity tasks and analogy tasks. The results demonstrate that the proposed models can outperform state-of-the-art models significantly on both word and phrase representation learning.
A Deep Architecture for Semantic Matching with Multiple Positional Sentence Representations
Wan, Shengxian (Chinese Academy of Sciences) | Lan, Yanyan (Chinese Academy of Sciences) | Guo, Jiafeng (Chinese Academy of Sciences) | Xu, Jun (Chinese Academy of Sciences) | Pang, Liang (Chinese Academy of Sciences) | Cheng, Xueqi (Chinese Academy of Sciences)
Matching natural language sentences is central for many applications such as information retrieval and question answering. Existing deep models rely on a single sentence representation or multiple granularity representations for matching. However, such methods cannot well capture the contextualized local information in the matching process. To tackle this problem, we present a new deep architecture to match two sentences with multiple positional sentence representations. Specifically, each positional sentence representation is a sentence representation at this position, generated by a bidirectional long short term memory (Bi-LSTM). The matching score is finally produced by aggregating interactions between these different positional sentence representations, through k-Max pooling and a multi-layer perceptron. Our model has several advantages: (1) By using Bi-LSTM, rich context of the whole sentence is leveraged to capture the contextualized local information in each positional sentence representation; (2) By matching with multiple positional sentence representations, it is flexible to aggregate different important contextualized local information in a sentence to support the matching; (3) Experiments on different tasks such as question answering and sentence completion demonstrate the superiority of our model.
SPAN: Understanding a Question with Its Support Answers
Pang, Liang (Institute of Computing Technology, Chinese Academy of Sciences) | Lan, Yanyan (Institute of Computing Technology, Chinese Academy of Sciences) | Guo, Jiafeng (Institute of Computing Technology, Chinese Academy of Sciences) | Xu, Jun (Institute of Computing Technology, Chinese Academy of Sciences) | Cheng, Xueqi (Institute of Computing Technology, Chinese Academy of Sciences)
Matching a question to its best answer is a common task in community question answering. In this paper, we focus on the non-factoid questions and aim to pick out the best answer from its candidate answers. Most of the existing deep models directly measure the similarity between question and answer by their individual sentence embeddings. In order to tackle the problem of the information lack in question's descriptions and the lexical gap between questions and answers, we propose a novel deep architecture namely SPAN in this paper. Specifically we introduce support answers to help understand the question, which are defined as the best answers of those similar questions to the original one. Then we can obtain two kinds of similarities, one is between question and the candidate answer, and the other one is between support answers and the candidate answer. The matching score is finally generated by combining them. Experiments on Yahoo! Answers demonstrate that SPAN can outperform the baseline models.
Text Matching as Image Recognition
Pang, Liang (Chinese Academy of Sciences) | Lan, Yanyan (Chinese Academy of Sciences) | Guo, Jiafeng (Chinese Academy of Sciences) | Xu, Jun (Chinese Academy of Sciences) | Wan, Shengxian (Institute of Computing Technology, Chinese Academy of Sciences) | Cheng, Xueqi (Institute of Computing Technology, Chinese Academy of Sciences)
Matching two texts is a fundamental problem in many natural language processing tasks. An effective way is to extract meaningful matching patterns from words, phrases, and sentences to produce the matching score. Inspired by the success of convolutional neural network in image recognition, where neurons can capture many complicated patterns based on the extracted elementary visual patterns such as oriented edges and corners, we propose to model text matching as the problem of image recognition. Firstly, a matching matrix whose entries represent the similarities between words is constructed and viewed as an image. Then a convolutional neural network is utilized to capture rich matching patterns in a layer-by-layer way. We show that by resembling the compositional hierarchies of patterns in image recognition, our model can successfully identify salient signals such as n-gram and n-term matchings. Experimental results demonstrate its superiority against the baselines.
A Probabilistic Model for Bursty Topic Discovery in Microblogs
Yan, Xiaohui (Institute of Computing Technology, Chinese Academy of Science) | Guo, Jiafeng (Institute of Computing Technology, Chinese Academy of Science) | Lan, Yanyan (Institute of Computing Technology, Chinese Academy of Science) | Xu, Jun (Institute of Computing Technology, Chinese Academy of Science) | Cheng, Xueqi (Institute of Computing Technology, Chinese Academy of Science)
Bursty topics discovery in microblogs is important for people to grasp essential and valuable information. However, the task is challenging since microblog posts are particularly short and noisy. This work develops a novel probabilistic model, namely Bursty Biterm Topic Model (BBTM), to deal with the task. BBTM extends the Biterm Topic Model (BTM) by incorporating the burstiness of biterms as prior knowledge for bursty topic modeling, which enjoys the following merits: 1) It can well solve the data sparsity problem in topic modeling over short texts as the same as BTM; 2) It can automatical discover high quality bursty topics in microblogs in a principled and efficient way. Extensive experiments on a standard Twitter dataset show that our approach outperforms the state-of-the-art baselines significantly.