Title | ||
---|---|---|
Context-dependent Label Smoothing Regularization for Attention-based End-to-End Code-Switching Speech Recognition |
Abstract | ||
---|---|---|
Previous works utilize the context-independent (CI) label smoothing regularization (LSR) method to prevent attention-based End-to-End (E2E) automatic speech recognition (ASR) model, which is trained with a cross entropy loss function and hard labels, from making over-confident predictions. But the CI LSR method does not make use of linguistic knowledge within and between languages in the case of code-switching speech recognition (CSSR). In this paper, we propose the context-dependent (CD) LSR method. According to code-switching linguistic knowledge, the output units are classified into several categories and several context dependency rules are made. Under the guidance of the context dependency rules, prior label distribution is generated dynamically according to the category of historical context, rather than being fixed. Thus, the CD LSR method can utilize the linguistic knowledge in the case of CSSR to further improve the performance of the model. Experiments on the SEAME corpus demonstrate the effects of the proposed method. The final system with the CD LSR method achieves the best performance with 37.21% mixed error rate (MER), obtaining up to 3.7% relative MER reduction compared to the baseline system with no LSR method. |
Year | DOI | Venue |
---|---|---|
2021 | 10.1109/ISCSLP49672.2021.9362080 | 2021 12th International Symposium on Chinese Spoken Language Processing (ISCSLP) |
Keywords | DocType | ISBN |
speech recognition,code-switching,label smoothing regularization,attention-based End-to-End | Conference | 978-1-7281-6995-8 |
Citations | PageRank | References |
0 | 0.34 | 0 |
Authors | ||
5 |
Name | Order | Citations | PageRank |
---|---|---|---|
Zheying Huang | 1 | 0 | 0.34 |
Peng Li | 2 | 0 | 0.34 |
Ji Xu | 3 | 0 | 0.34 |
Pengyuan Zhang | 4 | 50 | 19.46 |
Yonghong Yan | 5 | 0 | 0.34 |