Back to Search
Start Over
Detecting and Isolating Adversarial Attacks Using Characteristics of the Surrogate Model Framework
- Source :
- Applied Sciences, Vol 13, Iss 17, p 9698 (2023)
- Publication Year :
- 2023
- Publisher :
- MDPI AG, 2023.
-
Abstract
- The paper introduces a novel framework for detecting adversarial attacks on machine learning models that classify tabular data. Its purpose is to provide a robust method for the monitoring and continuous auditing of machine learning models for the purpose of detecting malicious data alterations. The core of the framework is based on building machine learning classifiers for the detection of attacks and its type that operate on diagnostic attributes. These diagnostic attributes are obtained not from the original model, but from the surrogate model that has been created by observation of the original model inputs and outputs. The paper presents building blocks for the framework and tests its power for the detection and isolation of attacks in selected scenarios utilizing known attacks and public machine learning data sets. The obtained results pave the road for further experiments and the goal of developing classifiers that can be integrated into real-world scenarios, bolstering the robustness of machine learning applications.
Details
- Language :
- English
- ISSN :
- 20763417
- Volume :
- 13
- Issue :
- 17
- Database :
- Directory of Open Access Journals
- Journal :
- Applied Sciences
- Publication Type :
- Academic Journal
- Accession number :
- edsdoj.bca413c6e48540adb5daca1df09c3539
- Document Type :
- article
- Full Text :
- https://doi.org/10.3390/app13179698