Joint Learning of Word and Label Embeddings for Sequence Labelling in Spoken Language Understanding

Wu, Jiewen, D'Haro, Luis Fernando, Chen, Nancy F., Krishnaswamy, Pavitra, Banchs, Rafael E.

arXiv.org Artificial Intelligence 

Palo Alto, CA, 94306, USA ABSTRACT We propose an architecture to jointly learn word and label embeddings for slot filling in spoken language understanding. The proposed approach encodes labels using a combination of word embeddings and straightforward word-label association from the training data. Compared to the state-of- the-art methods, our approach does not require label embed-dings as part of the input and therefore lends itself nicely to a wide range of model architectures. In addition, our architecture computes contextual distances between words and labels to avoid adding contextual windows, thus reducing memory footprint. We validate the approach on established spoken dialogue datasets and show that it can achieve state-of-the-art performance with much fewer trainable parameters. Index T erms-- Slot-filling, recurrent neural network, distributional semantics, sequence labelling 1. INTRODUCTION In spoken language understanding (SLU), an essential step is to associate each word in an utterance with one semantic class label. These annotated utterances can then serve as a basis for higher level SLU tasks, such as topic identification and dialogue response generation. This process of semantic label tagging in SLU, dubbed slot filling, labels utterance sequences with tags under a specific scheme. As an example, the BIO scheme prefixes tags with one of the characters { B, I, O } to indicate the continuity of a tag: Begin, Inside, or Outside, e.g., B-price indicates this position is the beginning of the tag price. Researchers also developed deep learning architecture for slot filling, e.g., [1, 2, 3].

Duplicate Docs Excel Report

Title
None found

Similar Docs  Excel Report  more

TitleSimilaritySource
None found