Title
Combination Of Latent Semantic Analysis Based Language Models For Meeting Recognition
Abstract
Latent Semantic Analysis (LSA) defines a semantic similarity space using a training corpus. This semantic similarity can be used for dealing with long distance dependencies, which are an inherent problem for traditional wordbased n-gram models. Since LSA models adapt dynamically to topics, and meetings have clear topics, we conjecture that these models can improve speech recognition accuracy on meetings. This paper presents perplexity and word error rate results for LSA models for meetings. We present results for models trained on a variety of corpora including meeting data and background domain data, and for combinations of multiple LSA models together with a word-based n-gram model. We show that the meeting and background LSA models can improve over the baseline n-grain models in terms of perplexity and that some background LSA models can significantly improve over the n-gram models in terms of word error rate. For the combination of multiple LSA models we did however not see such an improvement.
Year
Venue
Keywords
2006
PROCEEDINGS OF THE SECOND IASTED INTERNATIONAL CONFERENCE ON COMPUTATIONAL INTELLIGENCE
speech recognition, latent semantic indexing
Field
DocType
Citations 
Perplexity,Computer science,Explicit semantic analysis,Document-term matrix,Artificial intelligence,Natural language processing,Probabilistic latent semantic analysis,Semantic computing,Semantic similarity,SemEval,Pattern recognition,Speech recognition,Latent semantic analysis,Machine learning
Conference
1
PageRank 
References 
Authors
0.35
8
3
Name
Order
Citations
PageRank
Michael Puscher110.35
Yan Huang210.35
Ozgur Cetin310.35