Title
Language-Invariant Bottleneck Features From Adversarial End-To-End Acoustic Models For Low Resource Speech Recognition
Abstract
This paper proposes to learn language-invariant bottleneck features from an adversarial end-to-end acoustic model for low resource languages. The multilingual end-to-end model is trained with a connectionist temporal classification loss function. The model has shared and private layers. The shared layers are the hidden layers utilized to learn universal features for all the languages. The private layers are the language-dependent layers used to capture language-specific features. Attention based adversarial end-to-end language identification is used to capture enough language information. Furthermore, orthogonality constraints are used to make private and shared features dissimilar. Experiments are conducted on IARPA Babel datasets. The results show that the target model trained with the proposed language-invariant bottleneck features outperforms the target model trained with the conventional multilingual bottleneck features by up to 9.7% relative word error rate reduction.
Year
DOI
Venue
2019
10.1109/icassp.2019.8682972
2019 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP)
Keywords
Field
DocType
Language-invariant, adversarial, end-to-end, low resource, speech recognition
Data modeling,Bottleneck,Computer science,Word error rate,Feature extraction,Speech recognition,Language identification,Hidden Markov model,Connectionism,Acoustic model
Conference
ISSN
Citations 
PageRank 
1520-6149
0
0.34
References 
Authors
0
3
Name
Order
Citations
PageRank
Jiangyan Yi11917.99
Jianhua Tao2848138.00
Ye Bai375.52