Title
A novel video annotation framework using near-duplicate segment detection
Abstract
The traditional video annotation approaches focus on annotating keyframes, shots, or the whole video with semantic keywords. However, the extractions of keyframes and shots lack of semantic meanings, and it is hard to use a few keywords to describe a video by using multiple topics. Therefore, we propose a novel video annotation framework using near-duplicate segment detection not only to preserve but also to purify the semantic meanings of target annotation units. A hierarchical near-duplicate segment detection method is proposed to efficiently localize near-duplicate segments in frame-level. Videos containing near-duplicate segments are clustered and keyword distributions of clusters are analyzed. Finally, the keywords ranked according to keyword distribution scores are annotated onto the obtained annotation units. Comprehensive experiments demonstrate the effectiveness of the proposed video annotation framework and near-duplicate segment detection method.
Year
DOI
Venue
2015
10.1109/ICMEW.2015.7169854
ICME Workshops
Keywords
Field
DocType
video annotation, automatic annotation, near-duplicate segment detection, web video analysis
Annotation,Ranking,Pattern recognition,Visualization,Computer science,Video annotation,Feature extraction,Redundancy (engineering),Artificial intelligence,Semantics
Conference
ISSN
Citations 
PageRank 
2330-7927
0
0.34
References 
Authors
10
4
Name
Order
Citations
PageRank
Chien-Li Chou18610.09
Hua-Tsung Chen228928.72
Chun-Chieh Hsu3165.22
Suh-Yin Lee41596319.67