Title
Neural-FST Class Language Model for End-to-End Speech Recognition
Abstract
We propose Neural-FST Class Language Model (NFCLM) for end-to-end speech recognition, a novel method that combines neural network language models (NNLMs) and finite state transducers (FSTs) in a mathematically consistent framework. Our method utilizes a background NNLM which models generic background text together with a collection of domain-specific entities modeled as individual FSTs. Each output token is generated by a mixture of these components; the mixture weights are estimated with a separately trained neural decider. We show that NFCLM significantly outperforms NNLM by 15.8% relative in terms of Word Error Rate. NFCLM achieves similar performance as traditional NNLM and FST shallow fusion while being less prone to overbiasing and 12 times more compact, making it more suitable for on-device usage.
Year
DOI
Venue
2022
10.1109/ICASSP43922.2022.9747573
IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP)
DocType
Citations 
PageRank 
Conference
0
0.34
References 
Authors
0
10
Name
Order
Citations
PageRank
Antoine Bruguier100.34
Duc-Vinh Le24515.88
Rohit Prabhavalkar316322.56
Dangna Li400.34
Zhe Liu500.34
Bo Wang600.34
Eun Chang700.34
Fuchun Peng8137885.75
Ozlem Kalinli913.39
Michael L. Seltzer1000.34