Back to Search Start Over

Diversity Adversarial Training against Adversarial Attack on Deep Neural Networks

Authors :
Hyun Kwon
Jun Lee
Source :
Symmetry, Vol 13, Iss 428, p 428 (2021), Symmetry; Volume 13; Issue 3; Pages: 428
Publication Year :
2021
Publisher :
MDPI AG, 2021.

Abstract

This paper presents research focusing on visualization and pattern recognition based on computer science. Although deep neural networks demonstrate satisfactory performance regarding image and voice recognition, as well as pattern analysis and intrusion detection, they exhibit inferior performance towards adversarial examples. Noise introduction, to some degree, to the original data could lead adversarial examples to be misclassified by deep neural networks, even though they can still be deemed as normal by humans. In this paper, a robust diversity adversarial training method against adversarial attacks was demonstrated. In this approach, the target model is more robust to unknown adversarial examples, as it trains various adversarial samples. During the experiment, Tensorflow was employed as our deep learning framework, while MNIST and Fashion-MNIST were used as experimental datasets. Results revealed that the diversity training method has lowered the attack success rate by an average of 27.2 and 24.3% for various adversarial examples, while maintaining the 98.7 and 91.5% accuracy rates regarding the original data of MNIST and Fashion-MNIST.

Details

Language :
English
ISSN :
20738994
Volume :
13
Issue :
428
Database :
OpenAIRE
Journal :
Symmetry
Accession number :
edsair.doi.dedup.....ae63552514b2a62816f7a941f421aed8