Title
Using Deep Time Delay Neural Network for Slot Filling in Spoken Language Understanding.
Abstract
Modeling the context of a target word is of fundamental importance in predicting the semantic label for slot filling task in Spoken Language Understanding (SLU). Although Recurrent Neural Network (RNN) has shown to successfully achieve the state-of-the-art results for SLU, and Bidirectional RNN is capable of obtaining further improvement by modeling information not only from the past, but also from the future, they only consider limited contextual information of the target word. In order to make the network deeper and hence obtain longer contextual information, we propose to use a multi-layer Time Delay Neural Network (TDNN), which is prevalent in current large vocabulary continuous speech recognition tasks. In particular, we use a TDNN with symmetric time delay offset. To make the stacked TDNN easily trained, residual structures and skip concatenation are adopted. In addition, we further improve the model by introducing ResTDNN-BiLSTM, which combines the advantages of both the residual TDNN and BiLSTM. Experiments on slot filling tasks on the Air Travel Information System (ATIS) and Snips benchmark datasets show the proposed SC-TDNN-C achieves state-of-the-art results without any additional knowledge and data resources. Finally, we review and compare slot filling results by using a variety of existing models and methods.
Year
DOI
Venue
2020
10.3390/sym12060993
SYMMETRY-BASEL
Keywords
DocType
Volume
Spoken Language Understanding,Time Delay Neural Network,residual network,skip concatenation
Journal
12
Issue
Citations 
PageRank 
6
0
0.34
References 
Authors
0
3
Name
Order
Citations
PageRank
Zhen Zhang101.01
Hao Huang261.93
Kai Wang300.34