Title
Human And Automated Scoring Of Fluency, Pronunciation And Intonation During Human-Machine Spoken Dialog Interactions
Abstract
We present a spoken dialog-based framework for the computer-assisted language learning (CALL) of conversational English. In particular, we leveraged the open-source HALEF dialog framework to develop a job interview conversational application. We then used crowdsourcing to collect multiple interactions with the system from non-native English speakers. We analyzed human-rated scores of the recorded dialog data on three different scoring dimensions critical to the delivery of conversational English - fluency, pronunciation and intonation/stress - and further examined the efficacy of automatically-extracted, hand-curated speech features in predicting each of these sub-scores. Machine learning experiments showed that trained scoring models generally perform at par with the human inter-rater agreement baseline in predicting human-rated scores of conversational proficiency.
Year
DOI
Venue
2017
10.21437/Interspeech.2017-1213
18TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2017), VOLS 1-6: SITUATED INTERACTION
Keywords
Field
DocType
dialog systems, computer assisted language learning, conversational assessment, intelligent tutoring systems, crowdsourcing
Pronunciation,Human–machine system,Spoken dialog,Computer science,Fluency,Speech recognition
Conference
ISSN
Citations 
PageRank 
2308-457X
0
0.34
References 
Authors
7
5
Name
Order
Citations
PageRank
Vikram Ramanarayanan17013.97
Patrick Lange298.42
Keelan Evanini314.42
Hillary R. Molloy400.34
David Suendermann-Oeft592.63