Title
2D character animating networks - bringing static characters to move via motion transfer.
Abstract
Animating a character given an image is an interesting research problem that posed to have applicational values. Namely, it could help animators to automatically animate new characters with motions similar to previous characters. The problem also has value in digital entertainment for animating characters created by players such as drawn characters. There are a few deep learning video synthesis models that could generate a video given an image. However, it is unlikely there exists a dataset that could train them for animating characters since characters for digital entertainment tend to be more novel and diverse. Nor is it practical for animators to create a large dataset for training. To this end, a 2D-CharAnimNet is proposed. Empowered by a novel motion transfer scheme for video generation, the proposed variational-autoencoder-based model could use a relatively small dataset. In addition, to improve the fidelity of video frames, dynamic skip-connections along with a polishing generative adversarial networks are also proposed. Results seem to indicate that the model has encouraging potential in adapting for applicational uses.
Year
DOI
Venue
2019
10.1145/3297280.3297301
SAC
Keywords
Field
DocType
animation, generative model, motion transfer, neural networks, video synthesis
Fidelity,Existential quantification,Computer science,Motion transfer,Human–computer interaction,Animation,Artificial intelligence,Generative grammar,Deep learning,Artificial neural network,Generative model
Conference
ISBN
Citations 
PageRank 
978-1-4503-5933-7
0
0.34
References 
Authors
0
5
Name
Order
Citations
PageRank
Zackary P. T. Sin102.03
Peter H. Ng2184.14
Simon Chi Keung Shiu366735.42
Fu-lai Chung424434.50
Hong Va Leong51099173.04