Title | ||
---|---|---|
Semantic video annotation by mining association patterns from visual and speech features |
Abstract | ||
---|---|---|
In this paper, we propose a novel approach for semantic video annotation through integrating visual features and speech features. By employing statistics and association patterns, the relations between video shots and human concept can be discovered effectively to conceptualize videos. In other words, the utilization of high-level rules can effectively complement the insufficiency of statistics-based methods in dealing with broad and complex keyword identification in video annotation. Empirical evaluations on NIST TRECVID video datasets reveal that our proposed approach can enhance the annotation accuracy substantially. |
Year | DOI | Venue |
---|---|---|
2008 | 10.1007/978-3-540-68125-0_110 | PAKDD |
Keywords | Field | DocType |
mining association pattern,novel approach,association pattern,semantic video annotation,video annotation,nist trecvid video datasets,speech feature,complex keyword identification,annotation accuracy,empirical evaluation,video shot | Annotation,Information retrieval,TRECVID,Computer science,Video annotation,NIST,Association rule learning,Dynamic Bayesian network | Conference |
Volume | ISSN | ISBN |
5012 | 0302-9743 | 3-540-68124-8 |
Citations | PageRank | References |
3 | 0.42 | 7 |
Authors | ||
4 |
Name | Order | Citations | PageRank |
---|---|---|---|
Vincent S. Tseng | 1 | 2923 | 161.33 |
Ja-Hwung Su | 2 | 329 | 24.53 |
Jhih-Hong Huang | 3 | 32 | 1.51 |
Chih-Jen Chen | 4 | 35 | 2.40 |