Abstract | ||
---|---|---|
While the technology of Deep Learning (DL) is a powerful tool when properly trained for image analysis and classification applications, some factors for its optimization rely solely on the training data and their environment. In an effort to tackle the problem of knowledge bias created during the training process of a Deep Neural Network (DNN) and specifically Adversarial Networks for image augmentation, this work presents an entirely unsupervised methodology for discovering the unfairness level of Deep Learning (DL) models and in extend, its wrongly accumulated or biased classes. Fdi, the proposed evaluation metric for quantizing the level of unfairness of a model is introduced, along with the method of weighting the model's knowledge and producing its weakest aspects in a data-agnostic way. |
Year | DOI | Venue |
---|---|---|
2021 | 10.1145/3465481.3470478 | ARES 2021: 16TH INTERNATIONAL CONFERENCE ON AVAILABILITY, RELIABILITY AND SECURITY |
Keywords | DocType | Citations |
Fairness, Adversarial networks, Federated Learning, Image synthesis, Image classification | Conference | 0 |
PageRank | References | Authors |
0.34 | 0 | 5 |
Name | Order | Citations | PageRank |
---|---|---|---|
Ilias Siniosoglou | 1 | 0 | 0.34 |
Vasileios Argyriou | 2 | 279 | 30.51 |
Stamatia Bibi | 3 | 0 | 0.34 |
Thomas Lagkas | 4 | 0 | 0.34 |
Panagiotis Sarigiannidis | 5 | 0 | 0.34 |