Abstract | ||
---|---|---|
We present a speech recognition system for the medical domain whose architecture is based on a state-of-the-art stack trained on over 270 h of medical speech data and 30 million tokens of text from clinical episodes. Despite the acoustic challenges and linguistic complexity of the domain, we were able to reduce the system’s word error rate to below 16% in a realistic clinical use case. To further benchmark our system, we determined the human word error rate on a corpus covering a wide variety of speakers, working with multiple medical transcriptionists, and found that our speech recognition system performs on a par with humans. |
Year | Venue | Field |
---|---|---|
2017 | SPECOM | Architecture,Computer science,Word error rate,Speech recognition,Linguistic sequence complexity,Parity (mathematics) |
DocType | Citations | PageRank |
Conference | 4 | 0.48 |
References | Authors | |
30 | 7 |
Name | Order | Citations | PageRank |
---|---|---|---|
Erik Edwards | 1 | 10 | 2.94 |
Wael Salloum | 2 | 59 | 6.86 |
Greg Finley | 3 | 8 | 1.88 |
James Fone | 4 | 7 | 1.20 |
Greg Cardiff | 5 | 4 | 0.48 |
Mark Miller | 6 | 10 | 3.96 |
David Suendermann-Oeft | 7 | 10 | 3.96 |