Title
A Novel Deep Multi-Modal Feature Fusion Method for Celebrity Video Identification
Abstract
In this paper, we develop a novel multi-modal feature fusion method for the 2019 iQIYI Celebrity Video Identification Challenge, which is held in conjunction with ACM MM 2019. The purpose of this challenge is to retrieve all the video clips of a given identity in the testing set. In this challenge, the multi-modal features of a celebrity are encouraged to be combined for a promising performance, such as face features, head features, body features, and audio features. As we know, the features from different modalities usually have their own influences on the results. To achieve better results, a novel weighted multi-modal feature fusion method is designed to obtain the final feature representation. After many experimental verification, we found that different feature fusion weights for training and testing make the method robust to multi-modal person identification. Experiments on the iQIYI-VID-2019 dataset show that our multi-modal feature fusion strategy effectively improves the accuracy of person identification. Specifically, for competition, we use a single model to get the result of 0.8952 in mAP, which ranks TOP-5 among all the competitive results.
Year
DOI
Venue
2019
10.1145/3343031.3356067
Proceedings of the 27th ACM International Conference on Multimedia
Keywords
Field
DocType
multi-modal feature fusion, person identification, video identification
Computer vision,Feature fusion,Computer science,Artificial intelligence,Modal
Conference
ISBN
Citations 
PageRank 
978-1-4503-6889-6
1
0.35
References 
Authors
0
6
Name
Order
Citations
PageRank
jianrong chen144.73
Li Yang213421.12
Yuanyuan Xu3265.58
Jing Huo49614.49
Yinghuan Shi520028.94
Yang Gao652850.36