Wu, Jiewen
Joint Learning of Word and Label Embeddings for Sequence Labelling in Spoken Language Understanding
Wu, Jiewen, D'Haro, Luis Fernando, Chen, Nancy F., Krishnaswamy, Pavitra, Banchs, Rafael E.
Palo Alto, CA, 94306, USA ABSTRACT We propose an architecture to jointly learn word and label embeddings for slot filling in spoken language understanding. The proposed approach encodes labels using a combination of word embeddings and straightforward word-label association from the training data. Compared to the state-of- the-art methods, our approach does not require label embed-dings as part of the input and therefore lends itself nicely to a wide range of model architectures. In addition, our architecture computes contextual distances between words and labels to avoid adding contextual windows, thus reducing memory footprint. We validate the approach on established spoken dialogue datasets and show that it can achieve state-of-the-art performance with much fewer trainable parameters. Index T erms-- Slot-filling, recurrent neural network, distributional semantics, sequence labelling 1. INTRODUCTION In spoken language understanding (SLU), an essential step is to associate each word in an utterance with one semantic class label. These annotated utterances can then serve as a basis for higher level SLU tasks, such as topic identification and dialogue response generation. This process of semantic label tagging in SLU, dubbed slot filling, labels utterance sequences with tags under a specific scheme. As an example, the BIO scheme prefixes tags with one of the characters { B, I, O } to indicate the continuity of a tag: Begin, Inside, or Outside, e.g., B-price indicates this position is the beginning of the tag price. Researchers also developed deep learning architecture for slot filling, e.g., [1, 2, 3].
Semantic Explanations of Predictions
Lecue, Freddy, Wu, Jiewen
The main objective of explanations is to transmit knowledge to humans. This work proposes to construct informative explanations for predictions made from machine learning models. Motivated by the observations from social sciences, our approach selects data points from the training sample that exhibit special characteristics crucial for explanation, for instance, ones contrastive to the classification prediction and ones representative of the models. Subsequently, semantic concepts are derived from the selected data points through the use of domain ontologies. These concepts are filtered and ranked to produce informative explanations that improves human understanding. The main features of our approach are that (1) knowledge about explanations is captured in the form of ontological concepts, (2) explanations include contrastive evidences in addition to normal evidences, and (3) explanations are user relevant.
Assertion Absorption in Object Queries over Knowledge Bases
Wu, Jiewen (University of Waterloo) | Hudek, Alexander (University of Waterloo) | Toman, David (University of Waterloo) | Weddell, Grant (University of Waterloo)
We develop a novel absorption technique for large collections of factual assertions about individual objects. These assertions are commonly accompanied by implicit background knowledge and form a knowledge base. Both the assertions and the background knowledge are expressed in a suitable language of Description Logic and queries over such knowledge bases can be expressed as assertion retrieval queries. The proposed absorption technique significantly improves the performance of such queries, in particular in cases where a large number of object features are known for the objects represented in such a knowledge base. In addition to the absorption technique we present the results of a preliminary experimental evaluation that validates the efficacy of the proposed optimization.