Abstract | ||
---|---|---|
The recent surge in interest for Deep Learning (motivated by its exceptional performances on longstanding problems) made Neural Networks a very appealing tool for many actors in our society. One issue in this shift of interest is that Neural Networks are very opaque objects and it is often hard to make sense of their predictions.
In this context, research efforts have focused on building fair representations of data which display little to no correlation with regard to a sensitive attribute s. In this paper we build onto a domain adaptation neural model by augmenting it with a "noise conditioning" mechanism which we show is instrumental in obtaining fair (i.e. non-correlated with s) representations. We provide experiments on standard datasets showing the effectiveness of the noise conditioning mechanism in helping the networks to ignore the sensible attribute.
|
Year | DOI | Venue |
---|---|---|
2020 | 10.1145/3341105.3374090 | SAC '20: The 35th ACM/SIGAPP Symposium on Applied Computing
Brno
Czech Republic
March, 2020 |
DocType | ISBN | Citations |
Conference | 978-1-4503-6866-7 | 1 |
PageRank | References | Authors |
0.35 | 0 | 3 |
Name | Order | Citations | PageRank |
---|---|---|---|
Mattia Cerrato | 1 | 1 | 0.35 |
Roberto Esposito | 2 | 64 | 10.87 |
Laura Li Puma | 3 | 1 | 0.35 |