Title | ||
---|---|---|
Find Objects And Focus On Highlights: Mining Object Semantics For Video Highlight Detection Via Graph Neural Networks |
Abstract | ||
---|---|---|
With the increasing prevalence of portable computing devices, browsing unedited videos is time-consuming and tedious. Video highlight detection has the potential to significantly ease this situation, which discoveries moments of user's major or special interest in a video. Existing methods suffer from two problems. Firstly, most existing approaches only focus on learning holistic visual representations of videos but ignore object semantics for inferring video highlights. Secondly, current state-of-the-art approaches often adopt the pairwise ranking-based strategy, which cannot enjoy the global information to infer highlights. Therefore, we propose a novel video highlight framework, named VH-GNN, to construct an object-aware graph and model the relationships between objects from a global view. To reduce computational cost, we decompose the whole graph into two types of graphs: a spatial graph to capture the complex interactions of object within each frame, and a temporal graph to obtain object-aware representation of each frame and capture the global information. In addition, we optimize the framework via a proposed multi-stage loss, where the first stage aims to determine the highlight-probability and the second stage leverage the relationships between frames and focus on hard examples from the former stage. Extensive experiments on two standard datasets strongly evidence that VH-GNN obtains significant performance compared with state-of-the-arts. |
Year | Venue | DocType |
---|---|---|
2020 | AAAI | Conference |
Volume | ISSN | Citations |
34 | 2159-5399 | 0 |
PageRank | References | Authors |
0.34 | 0 | 6 |
Name | Order | Citations | PageRank |
---|---|---|---|
Yingying Zhang | 1 | 7 | 1.94 |
Junyu Gao | 2 | 67 | 7.23 |
Xiaoshan Yang | 3 | 149 | 16.83 |
Chang Liu | 4 | 15 | 7.17 |
Yan Li | 5 | 32 | 7.53 |
Changsheng Xu | 6 | 4957 | 332.87 |