Title
Controlling biases and diversity in diverse image-to-image translation
Abstract
The task of unpaired image-to-image translation is highly challenging due to the lack of explicit cross-domain pairs of instances. We consider here diverse image translation (DIT), an even more challenging setting in which an image can have multiple plausible translations. This is normally achieved by explicitly disentangling content and style in the latent representation and sampling different styles codes while maintaining the image content. Despite the success of current DIT models, they are prone to suffer from bias. In this paper, we study the problem of bias in image-to-image translation. Biased datasets may add undesired changes (e.g. change gender or race in face images) to the output translations as a consequence of the particular underlying visual distribution in the target domain. In order to alleviate the effects of this problem we propose the use of semantic constraints that enforce the preservation of desired image properties. Our proposed model is a step towards unbiased diverse image-to-image translation (UDIT), and results in less unwanted changes in the translated images while still performing the wanted transformation. Experiments on several heavily biased datasets show the effectiveness of the proposed techniques in different domains such as faces, objects, and scenes.
Year
DOI
Venue
2021
10.1016/j.cviu.2020.103082
Computer Vision and Image Understanding
Keywords
DocType
Volume
41A05,41A10,65D05,65D17
Journal
202
Issue
ISSN
Citations 
1
1077-3142
1
PageRank 
References 
Authors
0.41
0
4
Name
Order
Citations
PageRank
Yaxing Wang185.25
Abel Gonzalez-Garcia2415.47
Luis Herranz319426.17
Joost van de Weijer42117124.82