Title
Federated synthetic data generation with differential privacy
Abstract
Distributed machine learning has attracted much attention in the last decade with the widespread use of the Internet of Things. As a generative model, Generative Adversarial Network (GAN) has excellent empirical performance. However, the distributed storage of data and the fact that data cannot be shared for privacy reasons in a federated learning setting bring new challenges to training GAN. To address this issue, we propose private FL-GAN, a differentially private GAN based on federated learning. By strategically combining the Lipschitz condition with differential privacy sensitivity, our model can generate high-quality synthetic data without sacrificing the training data’s privacy. When communication between clients becomes the main bottleneck for federated learning, we propose to use a serialized model-training paradigm, which significantly reduces communication costs. Considering the distributed data is often non-IID in reality, which poses challenges to modeling, we further propose universal private FL-GAN to approach this problem. We not only theoretically prove that our algorithms can provide strict privacy guarantees with differential privacy, but also experimentally demonstrate that our models can generate satisfactory data while protecting the privacy of the training data, even if the data is non-IID.
Year
DOI
Venue
2022
10.1016/j.neucom.2021.10.027
Neurocomputing
Keywords
DocType
Volume
Synthetic data,Generative adversarial network,Federated learning,Differential privacy
Journal
468
ISSN
Citations 
PageRank 
0925-2312
0
0.34
References 
Authors
0
7
Name
Order
Citations
PageRank
Bangzhou Xin101.35
Yangyang Geng231.40
Teng Hu300.34
Sheng Chen400.34
Wei Yang528654.48
Shaowei Wang6111985.65
Liusheng Huang7242.19