Title
Evaluating the Usability of Automatically Generated Captions for People who are Deaf or Hard of Hearing.
Abstract
The accuracy of Automated Speech Recognition (ASR) technology has improved, but it is still imperfect in many settings. Researchers who evaluate ASR performance often focus on improving the Word Error Rate (WER) metric, but WER has been found to have little correlation with human-subject performance on many applications. We propose a new captioning-focused evaluation metric that better predicts the impact of ASR recognition errors on the usability of automatically generated captions for people who are Deaf or Hard of Hearing (DHH). Through a user study with 30 DHH users, we compared our new metric with the traditional WER metric on a caption usability evaluation task. In a side-by-side comparison of pairs of ASR text output (with identical WER), the texts preferred by our new metric were preferred by DHH participants. Further, our metric had significantly higher correlation with DHH participants' subjective scores on the usability of a caption, as compared to the correlation between WER metric and participant subjective scores. This new metric could be used to select ASR systems for captioning applications, and it may be a better metric for ASR researchers to consider when optimizing ASR systems.
Year
DOI
Venue
2017
10.1145/3132525.3132542
ASSETS
Keywords
DocType
Volume
Accessibility for People who are Deaf or Hard-of-Hearing, Automatic Speech Recognition, Real-time Captioning System, Caption Usability Evaluation
Journal
abs/1712.02033
ISSN
ISBN
Citations 
ASSETS'17 (2017) 165-174
978-1-4503-4926-0
3
PageRank 
References 
Authors
0.41
21
2
Name
Order
Citations
PageRank
Sushant Kafle1104.03
Matt Huenerfauth242851.83