Title
MLPrivacyGuard: Defeating Confidence Information based Model Inversion Attacks on Machine Learning Systems
Abstract
As services based on Machine Learning (ML) applications find increasing use, there is a growing risk of attack against such systems. Recently, adversarial machine learning has received a lot of attention, where an adversary is able to craft an input or manipulate an input to cause an ML system to misclassify. Another attack of concern is when an adversary with access to a ML model can reverse engineer attributes of a target class, creating a privacy concern, which is the subject of this paper. Such attacks use non-sensitive data obtainable by the adversary and the confidence levels returned by the ML model to infer sensitive attributes of the target user. Model Inversion attacks may be classified as white-box, where the ML model is known to the attacker, or black-box, where the adversary does not know the internals of the model. If the attacker has access to non-sensitive data of a target user, they can infer sensitive data by applying gradient ascent on the confidence returned by the model. Therefore, a black-box attack can be mounted by numerical approximations of the gradient to perform the gradient ascent. In this work, we present MLPrivacyGuard, a countermeasure against black-box model inversion attack is presented. This countermeasure consists of adding controlled noise to the output of the confidence function. It is important to preserve the accuracy of prediction/classification for the real users of the model while preventing attackers to infer sensitive data. This involves a trade-off between misclassification error and the effectiveness of defense. Based on experimental results, we demonstrate that when noise is injected with a long-tailed distribution, the objectives of low misclassification error with a strong defense can be attained as model inversion attacks are neutralized because numerical approximation of gradient ascent is unable to converge.
Year
DOI
Venue
2019
10.1145/3299874.3319457
Proceedings of the 2019 on Great Lakes Symposium on VLSI
Keywords
Field
DocType
adversarial machine learning, convolutional neural network, model inversion attack, neural networks
Countermeasure,Gradient descent,Model inversion,Convolutional neural network,Computer science,Reverse engineering,Adversarial machine learning,Artificial intelligence,Adversary,Artificial neural network,Machine learning
Conference
ISSN
ISBN
Citations 
1066-1395
978-1-4503-6252-8
0
PageRank 
References 
Authors
0.34
0
3
Name
Order
Citations
PageRank
Tiago A. O. Alves1458.34
Felipe M. G. França224951.12
Sandip Kundu31103137.18