Abstract | ||
---|---|---|
Discriminatively trained HMMs are investigated in both clean and noisy environments in this study. First, a recognition error is defined at different levels including string, word, phone and acoustics. A high resolution error measure in terms of minimum divergence (MD) is specifically proposed and investigated along with other error measures. Using two speaker-independent continuous digit databases, Aurora2(English) and CNDigits (Mandarin Chinese), the recognition performance of recognizers, which are trained in terms of different error measures and using different training modes, is evaluated under different noise and SNR conditions. Experimental results show that discriminatively trained models performed better than the maximum likelihood baseline systems. Specifically, for MD trained systems, relative error reductions of 17.62% and 18.52% were obtained applying multi-training on Aurora2 and CNDigits, respectively. |
Year | DOI | Venue |
---|---|---|
2006 | 10.1007/11939993_39 | ISCSLP |
Keywords | Field | DocType |
error measure,md trained system,different error measure,discriminative hmms,different level,noisy speech recognition performance,different noise,relative error reduction,recognition error,discriminatively trained hmms,high resolution error measure,different training mode,high resolution,relative error,speech recognition,mandarin chinese,maximum likelihood | Speech processing,Pattern recognition,Computer science,Markov model,Signal-to-noise ratio,Speech recognition,Artificial intelligence,Hidden Markov model,String (computer science),Discriminative model,Mandarin Chinese,Approximation error | Conference |
Volume | ISSN | ISBN |
4274 | 0302-9743 | 3-540-49665-3 |
Citations | PageRank | References |
1 | 0.35 | 9 |
Authors | ||
5 |
Name | Order | Citations | PageRank |
---|---|---|---|
Jun Du | 1 | 2 | 1.11 |
Peng Liu | 2 | 38 | 5.69 |
Frank K. Soong | 3 | 1395 | 268.29 |
Jian-Lai Zhou | 4 | 184 | 20.85 |
Ren-Hua Wang | 5 | 344 | 41.36 |