Title
Improving Attention-Based End-to-End ASR Systems with Sequence-Based Loss Functions.
Abstract
Acoustic model and language model (LM) have been two major components in conventional speech recognition systems. They are normally trained independently, but recently there has been a trend to optimize both components simultaneously in a unified end-to-end (E2E) framework. However, the performance gap between the E2E systems and the traditional hybrid systems suggests that some knowledge has not yet been fully utilized in the new framework. An observation is that the current attention-based E2E systems could produce better recognition results when decoded with LMs which are independently trained with the same resource.In this paper, we focus on how to improve attention-based E2E systems without increasing model complexity or resorting to extra data. A novel training strategy is proposed for multi-task training with the connectionist temporal classification (CTC) loss. The sequence-based minimum Bayes risk (MBR) loss is also investigated. Our experiments on SWB 300hrs showed that both loss functions could significantly improve the baseline model performance. The additional gain from joint-LM decoding remains the same for CTC trained model but is only marginal for MBR trained model. This implies that while CTC loss function is able to capture more acoustic knowledge, MBR loss function exploits more word/character dependency.
Year
DOI
Venue
2018
10.1109/SLT.2018.8639587
SLT
Keywords
Field
DocType
Decoding,Training,Acoustics,Speech recognition,Data models,Task analysis,Transforms
Data modeling,Computer science,End-to-end principle,Speech recognition,Decoding methods,Hybrid system,Language model,Connectionism,Bayes' theorem,Acoustic model
Conference
ISSN
ISBN
Citations 
2639-5479
978-1-5386-4334-1
2
PageRank 
References 
Authors
0.39
0
8
Name
Order
Citations
PageRank
Jia Cui162.80
Chao Weng211319.75
Guangsen Wang3284.98
Jun Wang49228736.82
Wang Peidong584.60
Chengzhu Yu6163.77
Dan Su76420.89
Dong Yu86264475.73