Title
Semantic video annotation by mining association patterns from visual and speech features
Abstract
In this paper, we propose a novel approach for semantic video annotation through integrating visual features and speech features. By employing statistics and association patterns, the relations between video shots and human concept can be discovered effectively to conceptualize videos. In other words, the utilization of high-level rules can effectively complement the insufficiency of statistics-based methods in dealing with broad and complex keyword identification in video annotation. Empirical evaluations on NIST TRECVID video datasets reveal that our proposed approach can enhance the annotation accuracy substantially.
Year
DOI
Venue
2008
10.1007/978-3-540-68125-0_110
PAKDD
Keywords
Field
DocType
mining association pattern,novel approach,association pattern,semantic video annotation,video annotation,nist trecvid video datasets,speech feature,complex keyword identification,annotation accuracy,empirical evaluation,video shot
Annotation,Information retrieval,TRECVID,Computer science,Video annotation,NIST,Association rule learning,Dynamic Bayesian network
Conference
Volume
ISSN
ISBN
5012
0302-9743
3-540-68124-8
Citations 
PageRank 
References 
3
0.42
7
Authors
4
Name
Order
Citations
PageRank
Vincent S. Tseng12923161.33
Ja-Hwung Su232924.53
Jhih-Hong Huang3321.51
Chih-Jen Chen4352.40