Title
Context-dependent Label Smoothing Regularization for Attention-based End-to-End Code-Switching Speech Recognition
Abstract
Previous works utilize the context-independent (CI) label smoothing regularization (LSR) method to prevent attention-based End-to-End (E2E) automatic speech recognition (ASR) model, which is trained with a cross entropy loss function and hard labels, from making over-confident predictions. But the CI LSR method does not make use of linguistic knowledge within and between languages in the case of code-switching speech recognition (CSSR). In this paper, we propose the context-dependent (CD) LSR method. According to code-switching linguistic knowledge, the output units are classified into several categories and several context dependency rules are made. Under the guidance of the context dependency rules, prior label distribution is generated dynamically according to the category of historical context, rather than being fixed. Thus, the CD LSR method can utilize the linguistic knowledge in the case of CSSR to further improve the performance of the model. Experiments on the SEAME corpus demonstrate the effects of the proposed method. The final system with the CD LSR method achieves the best performance with 37.21% mixed error rate (MER), obtaining up to 3.7% relative MER reduction compared to the baseline system with no LSR method.
Year
DOI
Venue
2021
10.1109/ISCSLP49672.2021.9362080
2021 12th International Symposium on Chinese Spoken Language Processing (ISCSLP)
Keywords
DocType
ISBN
speech recognition,code-switching,label smoothing regularization,attention-based End-to-End
Conference
978-1-7281-6995-8
Citations 
PageRank 
References 
0
0.34
0
Authors
5
Name
Order
Citations
PageRank
Zheying Huang100.34
Peng Li200.34
Ji Xu300.34
Pengyuan Zhang45019.46
Yonghong Yan500.34