Abstract | ||
---|---|---|
Recent advances in image-to-image translation have led to some ways to generate multiple domain images through a single network. However, there is still a limit in creating an image of a target domain without a dataset on it. We propose a method that expands the concept of 'multidomain' from data to the loss area and learns the combined characteristics of each domain to dynamically infer translations of images in mixed domains. First, we introduce Sym-parameter and its learning method for variously mixed losses while synchronizing them with input conditions. Then, we propose Sym-parameterized Generative Network (SGN) which is empirically confirmed of learning mixed characteristics of various data and losses, and translating images to any mixed-domain without ground truths, such as 30% Van Gogh and 20% Monet and 40% snowy. |
Year | DOI | Venue |
---|---|---|
2019 | 10.1109/ICCV.2019.00490 | 2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019) |
Field | DocType | Volume |
Image translation,Computer vision,Parameterized complexity,Pattern recognition,Computer science,Inference,Artificial intelligence | Conference | 2019 |
Issue | ISSN | Citations |
1 | 1550-5499 | 0 |
PageRank | References | Authors |
0.34 | 0 | 4 |
Name | Order | Citations | PageRank |
---|---|---|---|
Simyung Chang | 1 | 2 | 6.15 |
Seonguk Park | 2 | 0 | 2.03 |
John Yang | 3 | 1 | 3.06 |
Nojun Kwak | 4 | 862 | 63.79 |