16th Annual Conference of the International Speech Communication Association

Dresden, Germany
September 6-10, 2015

Zero-Shot Semantic Parser for Spoken Language Understanding

Emmanuel Ferreira, Bassam Jabaian, Fabrice Lefèvre

LIA, France

Machine learning algorithms are now common in the state-of-the-art spoken language understanding models. But to reach good performance they must be trained on a potentially large amount of data which are not available for a variety of tasks and languages of interest. In this work, we present a novel zero-shot learning method, based on word embeddings, allowing to derive a full semantic parser for spoken language understanding.
    No annotated in-context data are needed, the ontological description of the target domain and generic word embedding features (learned from freely available general domain data) suffice to derive the model. Two versions are studied with respect to how the model parameters and decoding step are handled, including an extension of the proposed approach in the context of conditional random fields. We show that this model, with very little supervision, can reach instantly performance comparable to those obtained by either state-of-the-art carefully handcrafted rule-based or trained statistical models for extraction of dialog acts on the Dialog State Tracking test datasets (DSTC2 and 3).

Full Paper

Bibliographic reference.  Ferreira, Emmanuel / Jabaian, Bassam / Lefèvre, Fabrice (2015): "Zero-shot semantic parser for spoken language understanding", In INTERSPEECH-2015, 1403-1407.