Title
An Interactive Video Annotation Frameowrk with Multiple Modalities
Abstract
Active learning and semi-supervised learning methods are frequently applied in multimedia annotation tasks in order to reduce human labeling effort. However, in most of these methods only single modality is applied. This paper presents an interactive video annotation framework, which is based on semi-supervised learning and active learning with multiple multimodalities. In the proposed framework, unlabeled samples are iteratively selected to be annotated manually according to certain strategy which has taken the potentials of different modalities into account, and then a graph-based semi-supervised learning algorithm is conducted on each modality. This process repeats for several rounds, and the results obtained from multiple modalities are then fused to generate final output. The proposed framework is computationally efficient, and the experimental results on TRECVID 2005 benchmark show that the proposed framework considerably outperforms previous approaches.
Year
DOI
Venue
2007
10.1109/ICASSP.2007.366068
ICASSP (1)
Keywords
Field
DocType
trecvid 2005 benchmark,video annotation,video signal processing,learning (artificial intelligence),multimodality,active learning,graph theory,interactive video annotation framework,semisupervised learning methods,multiple modalities,degradation,learning artificial intelligence,training data,labeling,semi supervised learning
Interactive video,Graph theory,Modalities,Multimodality,Annotation,Active learning,Pattern recognition,Multiple modalities,TRECVID,Computer science,Artificial intelligence,Machine learning
Conference
Volume
ISSN
ISBN
1
1520-6149
1-4244-0727-3
Citations 
PageRank 
References 
0
0.34
4
Authors
5
Name
Order
Citations
PageRank
Meng Wang1453.19
Xian-Sheng Hua26566328.17
Yan Song373451.98
Li-Rong Dai41070117.92
Ren-Hua Wang534441.36