Abstract | ||
---|---|---|
In this paper, we present a new approach for classifying video content into semantic classes at a high level of abstraction by exploiting the connoted visual code. The method is based on the concept of supervised learning algorithms that have already been applied for the classification of written text and spoken language quite successfully. In order to extent this approach for classifying video content, a visual analog to words is constructed from signal-level visual features. A common bag-of-words approach is applied in order to represent video documents. Subsequently, support vector machines are trained to categorize the documents into known classes by using the proposed visual words. Experimental results indicating the classification performance are given and discussed. |
Year | DOI | Venue |
---|---|---|
2004 | 10.1016/j.cag.2004.03.002 | Computers & Graphics |
Keywords | DocType | Volume |
Video document classification,Video analysis,Supervised learning,Support vector machines,Content-based retrieval,Visual features | Journal | 28 |
Issue | ISSN | Citations |
3 | 0097-8493 | 0 |
PageRank | References | Authors |
0.34 | 3 | 5 |
Name | Order | Citations | PageRank |
---|---|---|---|
René Cavet | 1 | 0 | 0.34 |
Stephan Volmer | 2 | 57 | 3.74 |
Edda Leopold | 3 | 381 | 30.50 |
Jörg Kindermann | 4 | 411 | 33.66 |
Gerhard Paass | 5 | 1136 | 83.63 |