Title
Semi-Supervised Training Of Deep Neural Networks
Abstract
In this paper we search for an optimal strategy for semi-supervised Deep Neural Network (DNN) training. We assume that a small part of the data is transcribed, while the majority of the data is untranscribed. We explore self-training strategies with data selection based on both the utterance-level and frame-level confidences. Further on, we study the interactions between semi-supervised frame-discriminative training and sequence-discriminative sMBR training. We found it beneficial to reduce the disproportion in amounts of transcribed and untranscribed data by including the transcribed data several times, as well as to do a frame-selection based on per-frame confidences derived from confusion in a lattice.For the experiments, we used the Limited language pack condition for the Surprise language task (Vietnamese) from the IARPA Babel program. The absolute Word Error Rate (WER) improvement for frame cross-entropy training is 2.2%, this corresponds to WER recovery of 36% when compared to the identical system, where the DNN is built on the fully transcribed data.
Year
Venue
Keywords
2013
2013 IEEE WORKSHOP ON AUTOMATIC SPEECH RECOGNITION AND UNDERSTANDING (ASRU)
semi-supervised training, self-training, deep network, DNN, Babel program
Field
DocType
Citations 
Confusion,Data selection,Computer science,Word error rate,Speech recognition,Supervised training,Surprise,Artificial neural network,Deep neural networks
Conference
23
PageRank 
References 
Authors
1.08
13
3
Name
Order
Citations
PageRank
Karel Veselý115414.62
Mirko Hannemann21086.75
Lukas Burget3582.48