Title
Deep Transfer Learning for Recognizing Functional Interactions via Head Movements in Multiparty Conversations
Abstract
ABSTRACT Head movements play various functions in multiparty conversations. To date, convolutional neural networks (CNNs) have been proposed to recognize the functions of individual interlocutors’ head movements. This paper extends the concept of head-movement functions to the interaction functions between speaker and listener, which are performed through their head movements, e.g., a listener’s back-channel nodding in response to a speaker’s rhythmic movements. Then, we propose transfer strategies to build deep neural networks (DNNs) to recognize these interaction functions by reusing pretrained CNNs for individual head-movement functions. One of the proposed strategies uses CNNs as the feature extractor and identifies the interaction function with another classifier using the extracted features. Compared with the baseline model that employs the logical product of the output of two individual CNNs, the transferred DNNs outperform the baseline model in four out of five interaction functions. For example, the F-measure is improved by 13.9 points for the interaction of a listener’s positive emotion in response to a speaker’s rhythmic movements. These results confirm the potential of the proposed transfer strategies for recognizing interaction functions based on head movements.
Year
DOI
Venue
2021
10.1145/3462244.3479899
Multimodal Interfaces and Machine Learning for Multimodal Interaction
DocType
Citations 
PageRank 
Conference
0
0.34
References 
Authors
0
2
Name
Order
Citations
PageRank
Takashi Mori100.34
Kazuhiro Otsuka261954.15