Title
Deep Cg2real: Synthetic-To-Real Translation Via Image Disentanglement
Abstract
We present a method to improve the visual realism of low-quality, synthetic images, e.g. OpenGL renderings. Training an unpaired synthetic-to-real translation network in image space is severely under-constrained and produces visible artifacts. Instead, we propose a semi-supervised approach that operates on the disentangled shading and albedo layers of the image. Our two-stage pipeline first learns to predict accurate shading in a supervised fashion using physically-based renderings as targets, and further increases the realism of the textures and shading with an improved CycleGAN network. Extensive evaluations on the SUNCG indoor scene dataset demonstrate that our approach yields more realistic images compared to other state-of-the-art approaches. Furthermore, networks trained on our generated "real" images predict more accurate depth and normals than domain adaptation approaches, suggesting that improving the visual realism of the images can be more effective than imposing task-specific losses.
Year
DOI
Venue
2019
10.1109/ICCV.2019.00282
2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019)
Field
DocType
Volume
Computer vision,Computer science,Artificial intelligence
Conference
2019
Issue
ISSN
Citations 
1
1550-5499
1
PageRank 
References 
Authors
0.35
3
6
Name
Order
Citations
PageRank
Sai Bi1635.28
Kalyan Sunkavalli250031.75
Federico Perazzi326012.63
Eli Shechtman44340177.94
Vladimir G. Kim596141.44
Ravi Ramamoorthi64481237.21