Abstract | ||
---|---|---|
This paper introduces a novel deep learning framework for image animation. Given an input image with a target object and a driving video sequence depicting a moving object, our framework generates a video in which the target object is animated according to the driving sequence. This is achieved through a deep architecture that decouples appearance and motion information. Our framework consists of three main modules: (i) a Keypoint Detector unsupervisely trained to extract object keypoints, (ii) a Dense Motion prediction network for generating dense heatmaps from sparse keypoints, in order to better encode motion information and (iii) a Motion Transfer Network, which uses the motion heatmaps and appearance information extracted from the input image to synthesize the output frames. We demonstrate the effectiveness of our method on several benchmark datasets, spanning a wide variety of object appearances, and show that our approach outperforms state-of-the-art image animation and video generation methods. |
Year | DOI | Venue |
---|---|---|
2018 | 10.1109/CVPR.2019.00248 | 2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019) |
Field | DocType | Volume |
Computer vision,ENCODE,Computer science,Source code,Motion transfer,Animation,Artificial intelligence,Deep learning,Motion prediction,Detector | Journal | abs/1812.08861 |
ISSN | Citations | PageRank |
1063-6919 | 11 | 0.55 |
References | Authors | |
19 | 5 |
Name | Order | Citations | PageRank |
---|---|---|---|
Aliaksandr Siarohin | 1 | 28 | 3.06 |
Stéphane Lathuilière | 2 | 33 | 5.98 |
Sergey Tulyakov | 3 | 28 | 9.28 |
Elisa Ricci 0002 | 4 | 1393 | 73.75 |
Nicu Sebe | 5 | 7013 | 403.03 |