Affiliation:
1. University of South Florida, Tampa, FL
Abstract
Adversarial machine learning is a prominent research area aimed towards exposing and mitigating security vulnerabilities in AI/ML algorithms and their implementations. Data poisoning and neural Trojans enable an attacker to drastically change the behavior and performance of a
Convolutional Neural Network (CNN)
merely by altering some of the input data during training. Such attacks can be catastrophic in the field, e.g. for self-driving vehicles. In this paper, we propose deploying a CNN as an
ecosystem of variants
, rather than a singular model. The ecosystem is derived from the original trained model, and though every derived model is structurally different, they are all functionally equivalent to the original and each other. We propose two complementary techniques:
stochastic parameter mutation
, where the weights θ of the original are shifted by a small, random amount, and a
delta-update
procedure which functions by XOR’ing all of the parameters with an update file containing the Δ θ values. This technique is effective against transferability of a neural Trojan to the greater ecosystem by amplifying the Trojan’s malicious impact to easily detectable levels; thus, deploying a model as an ecosystem can render the ecosystem more resilient against a neural Trojan attack.
Publisher
Association for Computing Machinery (ACM)
Subject
Electrical and Electronic Engineering,Hardware and Architecture,Software
Reference61 articles.
1. Hacking smart machines with smarter ones: How to extract meaningful data from machine learning classifiers
2. Eugene Bagdasaryan Andreas Veit Yiqing Hua Deborah Estrin and Vitaly Shmatikov. 2019. How To Backdoor Federated Learning. arxiv:cs.CR/1807.00459
3. Detecting Poisoning Attacks on Machine Learning in IoT Environments
4. Randomized instruction set emulation to disrupt binary code injection attacks
5. Arjun Nitin Bhagoji Supriyo Chakraborty Prateek Mittal and Seraphin Calo. 2018. Analyzing Federated Learning through an Adversarial Lens. arxiv:cs.LG/1811.12470