Title | ||
---|---|---|
Combination Of Latent Semantic Analysis Based Language Models For Meeting Recognition |
Abstract | ||
---|---|---|
Latent Semantic Analysis (LSA) defines a semantic similarity space using a training corpus. This semantic similarity can be used for dealing with long distance dependencies, which are an inherent problem for traditional wordbased n-gram models. Since LSA models adapt dynamically to topics, and meetings have clear topics, we conjecture that these models can improve speech recognition accuracy on meetings. This paper presents perplexity and word error rate results for LSA models for meetings. We present results for models trained on a variety of corpora including meeting data and background domain data, and for combinations of multiple LSA models together with a word-based n-gram model. We show that the meeting and background LSA models can improve over the baseline n-grain models in terms of perplexity and that some background LSA models can significantly improve over the n-gram models in terms of word error rate. For the combination of multiple LSA models we did however not see such an improvement. |
Year | Venue | Keywords |
---|---|---|
2006 | PROCEEDINGS OF THE SECOND IASTED INTERNATIONAL CONFERENCE ON COMPUTATIONAL INTELLIGENCE | speech recognition, latent semantic indexing |
Field | DocType | Citations |
Perplexity,Computer science,Explicit semantic analysis,Document-term matrix,Artificial intelligence,Natural language processing,Probabilistic latent semantic analysis,Semantic computing,Semantic similarity,SemEval,Pattern recognition,Speech recognition,Latent semantic analysis,Machine learning | Conference | 1 |
PageRank | References | Authors |
0.35 | 8 | 3 |
Name | Order | Citations | PageRank |
---|---|---|---|
Michael Puscher | 1 | 1 | 0.35 |
Yan Huang | 2 | 1 | 0.35 |
Ozgur Cetin | 3 | 1 | 0.35 |