51. Adversarial Deep Learning with Stackelberg Games
- Author
-
Xinghao Yang, Aneesh Sreevallabh Chivukula, and Wei Liu
- Subjects
TheoryofComputation_MISCELLANEOUS ,Exploit ,Computer science ,business.industry ,Deep learning ,Stochastic game ,02 engineering and technology ,010501 environmental sciences ,Adversary ,Computer security ,computer.software_genre ,01 natural sciences ,Adversarial system ,symbols.namesake ,Incentive ,Nash equilibrium ,020204 information systems ,0202 electrical engineering, electronic engineering, information engineering ,Stackelberg competition ,symbols ,Artificial intelligence ,business ,Classifier (UML) ,computer ,0105 earth and related environmental sciences - Abstract
Deep networks are vulnerable to adversarial attacks from malicious adversaries. Currently, many adversarial learning algorithms are designed to exploit such vulnerabilities in deep networks. These methods focus on attacking and retraining deep networks with adversarial examples to do either feature manipulation or label manipulation or both. In this paper, we propose a new adversarial learning algorithm for finding adversarial manipulations to deep networks. We formulate adversaries who optimize game-theoretic payoff functions on deep networks doing multi-label classifications. We model the interactions between a classifier and an adversary from a game-theoretic perspective and formulate their strategies into a Stackelberg game associated with a two-player problem. Then we design algorithms to solve for the Nash equilibrium, which is a pair of strategies from which there is no incentive for either the classifier or the adversary to deviate. In designing attack scenarios, the adversary’s objective is to deliberately make small changes to test data such that attacked samples are undetected. Our results illustrate that game-theoretic modelling is significantly effective in securing deep learning models against performance vulnerabilities attached by intelligent adversaries.
- Published
- 2019