Title | ||
---|---|---|
A comparative study on methods of Weighted language model training for reranking lvcsr N-best hypotheses |
Abstract | ||
---|---|---|
This paper focuses on discriminative n-gram language models for a large vocabulary speech recognition task. Specifically we compare three training methods, Reranking Boosting (ReBst), Minimum Error Rate Training (MERT) and the Weighted Global Log-Linear Model (W-GCLM). They have a mechanism for handling sample weights, which are useful for providing an accurate model and work as impact factors of hypotheses for training. W-GCLM is proposed in this paper. We discuss the relationship between the three methods by comparing their loss functions. We also compare them experimentally by reranking N-best hypotheses under several conditions. We show that MERT and W-GCLM are different types of expansion of ReBst and have different respective advantages. Our experimental results reveal that W-GCLM outperforms ReBst and whether MERT or W-GCLM is superior depends on the training and test conditions. |
Year | DOI | Venue |
---|---|---|
2010 | 10.1109/ICASSP.2010.5495028 | Acoustics Speech and Signal Processing |
Keywords | Field | DocType |
speech recognition,N-best hypotheses reranking,discriminative n-gram language model,loss function,minimum error rate training,reranking boosting,vocabulary speech recognition task,weighted global log-linear model,weighted language model training,Discriminative LM,Error Correction,MERT,Reranking Boost,Weighted GCLM | Pattern recognition,Computer science,Word error rate,Natural language,Boosting (machine learning),Artificial intelligence,Estimation theory,Hidden Markov model,Vocabulary,Discriminative model,Language model,Machine learning | Conference |
ISSN | ISBN | Citations |
1520-6149 E-ISBN : 978-1-4244-4296-6 | 978-1-4244-4296-6 | 12 |
PageRank | References | Authors |
0.60 | 7 | 3 |
Name | Order | Citations | PageRank |
---|---|---|---|
Takanobu Oba | 1 | 53 | 12.09 |
Takaaki Hori | 2 | 408 | 45.58 |
Atsushi Nakamura | 3 | 57 | 13.11 |