An Understanding of the Vulnerability of Datasets to Disparate Membership Inference Attacks
-
Published:2022-12-14
Issue:4
Volume:2
Page:882-906
-
ISSN:2624-800X
-
Container-title:Journal of Cybersecurity and Privacy
-
language:en
-
Short-container-title:JCP
Author:
Moore Hunter D.ORCID, Stephens Andrew, Scherer WilliamORCID
Abstract
Recent efforts have shown that training data is not secured through the generalization and abstraction of algorithms. This vulnerability to the training data has been expressed through membership inference attacks that seek to discover the use of specific records within the training dataset of a model. Additionally, disparate membership inference attacks have been shown to achieve better accuracy compared with their macro attack counterparts. These disparate membership inference attacks use a pragmatic approach to attack individual, more vulnerable sub-sets of the data, such as underrepresented classes. While previous work in this field has explored model vulnerability to these attacks, this effort explores the vulnerability of datasets themselves to disparate membership inference attacks. This is accomplished through the development of a vulnerability-classification model that classifies datasets as vulnerable or secure to these attacks. To develop this model, a vulnerability-classification dataset is developed from over 100 datasets—including frequently cited datasets within the field. These datasets are described using a feature set of over 100 features and assigned labels developed from a combination of various modeling and attack strategies. By averaging the attack accuracy over 13 different modeling and attack strategies, the authors explore the vulnerabilities of the datasets themselves as opposed to a particular modeling or attack effort. The in-class observational distance, width ratio, and the proportion of discrete features are found to dominate the attributes defining dataset vulnerability to disparate membership inference attacks. These features are explored in deeper detail and used to develop exploratory methods for hardening these class-based sub-datasets against attacks showing preliminary mitigation success with combinations of feature reduction and class-balancing strategies.
Reference71 articles.
1. Algorithms that remember: Model inversion attacks and data protection law;Veale;Philos. Trans. R. Soc. A Math. Phys. Eng. Sci.,2018 2. Chakraborty, A., Alam, M., Dey, V., Chattopadhyay, A., and Mukhopadhyay, D. (2018). Adversarial attacks and defences: A survey. arXiv. 3. He, Y., Meng, G., Chen, K., Hu, X., and He, J. (2019). Towards Privacy and Security of Deep Learning Systems: A Survey. arXiv. 4. Qiu, S., Liu, Q., Zhou, S., and Wu, C. (2019). Review of artificial intelligence adversarial attack and defense technologies. Appl. Sci., 9. 5. Calandrino, J.A., Kilzer, A., Narayanan, A., Felten, E.W., and Shmatikov, V. (2011, January 22–25). “You might also like:” Privacy risks of collaborative filtering. Proceedings of the 2011 IEEE Symposium on Security and Privacy, Washington, DC, USA.
Cited by
3 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献
|
|