Abstract | ||
---|---|---|
Speech-reading is an invaluable technique for people with hearing loss or those in adverse listening conditions (e.g., in a noisy restaurant, near children playing loudly). However, speech-reading is often difficult because identical mouth shapes (visemes) can produce several speech sounds (phonemes); there is a one-to-many mapping from visemes to phonemes. This decreases comprehension, causing confusion and frustration during conversation. My doctoral research aims to design and evaluate a visualisation technique that displays textual representations of a speaker's phonemes to a speech-reader. By combining my visualisation with their pre-existing speech-reading ability, speech-readers should be able to disambiguate confusing viseme-to-phoneme mappings without shifting their focus from the speaker's face. This will result in an improved level of comprehension, supporting natural conversation.
|
Year | Venue | Field |
---|---|---|
2016 | ACM SIGACCESS | Visualisation technique,Speech sounds,Confusion,Conversation,Computer science,Viseme,Visualization,Active listening,Speech recognition,Artificial intelligence,Natural language processing,Comprehension |
DocType | Volume | Citations |
Journal | 114 | 0 |
PageRank | References | Authors |
0.34 | 0 | 1 |
Name | Order | Citations | PageRank |
---|---|---|---|
Benjamin M. Gorman | 1 | 7 | 2.59 |