Title
A salience-driven approach to speech recognition for human-robot interaction
Abstract
We present an implemented model for speech recognition in natural environments which relies on contextual information about salient entities to prime utterance recognition. The hypothesis underlying our approach is that, in situated human-robot interaction, speech recognition performance can be significantly enhanced by exploiting knowledge about the immediate physical environment and the dialogue history. To this end, visual salience (objects perceived in the physical scene) and linguistic salience (previously referred-to objects within the current dialogue) are integrated into a single cross-modal salience model. The model is dynamically updated as the environment evolves, and is used to establish expectations about uttered words which are most likely to be heard given the context. The update is realised by continously adapting the word-class probabilities specified in the statistical language model. The present article discusses the motivations behind our approach, describes our implementation as part of a distributed, cognitive architecture for mobile robots, and reports the evaluation results on a test suite.
Year
DOI
Venue
2009
10.1007/978-3-642-14729-6_8
ESSLLI Student Sessions
Keywords
Field
DocType
visual salience,dialogue history,speech recognition,speech recognition performance,human-robot interaction,environment evolves,salience-driven approach,prime utterance recognition,linguistic salience,single cross-modal salience model,current dialogue,statistical language model,cognitive architecture,human robot interaction,mobile robot
Computer science,Speech recognition,Cognitive model,Salience (language),Human–robot interaction
Conference
Volume
ISSN
ISBN
6211
0302-9743
3-642-14728-3
Citations 
PageRank 
References 
1
0.38
14
Authors
1
Name
Order
Citations
PageRank
Pierre Lison114612.35