Title
Simultaneous visual-appearance-level and spatial-temporal-level dictionary learning for video-based person re-identification
Abstract
Person re-identification (re-id) plays an important role in video surveillance and forensics applications. In many cases, person re-id should be conducted between video clips, i.e., given a query pedestrian video from one camera, the re-id system should retrieve the video clips containing the same person from other cameras. However, person re-id between videos, which we call video-based person re-id, has not been well studied. In this paper, we propose a visual-appearance-level and spatial-temporal-level dictionary learning (VSDL) approach for video-based person re-id. Specifically, we first employ two kinds of models to represent each walking cycle in the video, i.e., visual-appearance features of all frames within the walking cycle, and a spatial-temporal feature vector. By separately learning a visual-appearance-level dictionary and a spatial-temporal-level dictionary from two kinds of representations, each walking cycle can be represented as a coding coefficient. To enhance the discriminative ability of the obtained coding coefficients, we design a representation coefficient discriminant term for VSDL. Experiments on the public iLIDS-VID and PRID 2011 datasets demonstrate the effectiveness of VSDL.
Year
DOI
Venue
2019
10.1007/s00521-018-3529-7
Neural Computing and Applications
Keywords
Field
DocType
Video-based person re-identification, Dictionary learning, Spatial-temporal information, Visual-appearance information
Pedestrian,Feature vector,Dictionary learning,Discriminant,Speech recognition,Coding (social sciences),Artificial intelligence,Discriminative model,Machine learning,Mathematics,CLIPS,Visual appearance
Journal
Volume
Issue
ISSN
31
11
1433-3058
Citations 
PageRank 
References 
2
0.38
38
Authors
6
Name
Order
Citations
PageRank
Xiaoke Zhu1787.77
Xiaoke Zhu221.40
Xiao-Yuan Jing376955.18
Fei Ma45213.61
Li Cheng531.08
Yilin Ren620.38