Back to Search
Start Over
TextGuise: Adaptive adversarial example attacks on text classification model.
- Source :
-
Neurocomputing . Apr2023, Vol. 529, p190-203. 14p. - Publication Year :
- 2023
-
Abstract
- [Display omitted] • We propose a black-box adversarial text generation scheme called TextGuise, which can be used for almost all text classification tasks. The experiments show that TextGuise is superior to the state-of-the-art text adversarial example generation methods in terms of its attack success rate and attack efficiency when the perturbation ratio does not exceed 0.2.The adversarial examples generated using TextGuise retain the semantic information and strong readability of the text. None of the current defense methods can effectively defend against TextGuise. Adversarial examples greatly compromise the security of deep learning models. The key to improving the robustness of a natural language processing (NLP) model is to study attacks and defenses involving adversarial text. However, the current adversarial attack methods still face problems, such as the low success rates of attacks on some datasets, and the existing defense methods can already successfully defend against some attack methods. As a result, such attacks are unable to dig deeper into the flaws of NLP models to inform further defense improvements. Hence, it is necessary to design an adversarial attack method with a wider attack range and stronger performance. Aiming at the advantages and disadvantages of existing methods, this paper proposes a new adaptive black-box text adversarial example generation scheme, TextGuise. First, we design a keyword selection method in which word scores are calculated by combining context semantics to select the appropriate keywords to modify. Second, to maintain semantics, new keyword substitution rules are designed in combination with the characteristics of text and popular text expressions. Finally, the best modification strategy is adaptively selected through a querying model to reduce the magnitudes of disturbances. TextGuise can automatically select replacement keywords and replacement strategies that efficiently generate adversarial examples with good readability for various text classification tasks. Attack experiments conducted with TextGuise on 5 datasets yield high attack success rates that can surpass 80% when the perturbation ratio does not exceed 0.2. In addition, we present and discuss experiments focusing on defense, text similarity, query times, time consumption, etc., to test the attack performance of TextGuise. The results show that our attack method can achieve a good balance among various metrics. [ABSTRACT FROM AUTHOR]
- Subjects :
- *NATURAL language processing
*DEEP learning
*PLANT defenses
*CLASSIFICATION
Subjects
Details
- Language :
- English
- ISSN :
- 09252312
- Volume :
- 529
- Database :
- Academic Search Index
- Journal :
- Neurocomputing
- Publication Type :
- Academic Journal
- Accession number :
- 162061344
- Full Text :
- https://doi.org/10.1016/j.neucom.2023.01.071