1. AI-Q: a Framework to Measure the Robustness of Text Classification Models.
- Author
-
Mysliwiec, Konrad, Chinea-Ríos, Mara, Borrego-Obrador, Ian, and Franco-Salvador, Marc
- Subjects
DATA augmentation ,FRAGMENTED landscapes ,AREA studies ,CLASSIFICATION ,DATA modeling - Abstract
Robustness analysis of text Classification models through adversarial attacks has gained substantial attention in recent research. This area studies the consistent behavior of text Classification models under attacks. These attacks use perturbation methods based on applying semantic and label-preserving changes to the inputs. However, the fragmented landscape of individual attack implementations, dispersed across code repositories, poses complicates the development and application of comprehensive adversarial strategies for model enhancement. To address these challenges, this paper introduces AI-Q, a Python framework specifically designed for text Classification adversarial attacks and data augmentation. One of the major strengths of our framework lies in its extensive library of perturbation methods for adversarial attacks (24 in total), and its evaluation metrics for model robustness. The framework exhibits versatility by supporting both custom models and those from the HuggingFace ecosystem, ensuring broad compatibility with leading benchmarks in the field. Beyond adversarial attacks, AI-Q can be used for data augmentation, enabling users to harness the components of adversarial attacks to increase dataset diversity. Finally, our evaluation, including human annotations, highlights the AI-Q potential for model robustness. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF