Title
Fusion of Gesture and Speech for Increased Accuracy in Human Robot Interaction
Abstract
An approach for decision-level fusion for gesture and speech based human-robot interaction (HRI) is proposed. A rule-based method is compared with several machine learning approaches. Gestures and speech signals are initially classified using hidden Markov models, reaching accuracies of 89.6% and 84% respectively. The rule-based approach reached 91.6% while SVM, which was the best of all evaluated machine learning algorithms, reached an accuracy of 98.2% on the test data. A complete framework is deployed in real time humanoid robot (NAO) which proves the efficacy of the system.
Year
DOI
Venue
2019
10.1109/MMAR.2019.8864671
2019 24th International Conference on Methods and Models in Automation and Robotics (MMAR)
Keywords
Field
DocType
Gesture recognition,speech recognition,decision level fusion,HMM
Gesture,Computer science,Support vector machine,Gesture recognition,Speech recognition,Control engineering,Feature extraction,Test data,Hidden Markov model,Human–robot interaction,Humanoid robot
Conference
ISBN
Citations 
PageRank 
978-1-7281-0934-3
0
0.34
References 
Authors
8
3
Name
Order
Citations
PageRank
Neha Baranwal174.95
Avinash Kumar Singh23113.77
Thomas Hellström36410.98