Back to Search Start Over

Generation-driven Contrastive Self-training for Zero-shot Text Classification with Instruction-tuned GPT

Authors :
Zhang, Ruohong
Wang, Yau-Shian
Yang, Yiming
Publication Year :
2023

Abstract

Moreover, GPT-based zero-shot classification models tend to make independent predictions over test instances, which can be sub-optimal as the instance correlations and the decision boundaries in the target space are ignored. To address these difficulties and limitations, we propose a new approach to zero-shot text classification, namely \ourmodelshort, which leverages the strong generative power of GPT to assist in training a smaller, more adaptable, and efficient sentence encoder classifier with contrastive self-training. Specifically, GenCo applies GPT in two ways: firstly, it generates multiple augmented texts for each input instance to enhance the semantic embedding of the instance and improve the mapping to relevant labels; secondly, it generates augmented texts conditioned on the predicted label during self-training, which makes the generative process tailored to the decision boundaries in the target space. In our experiments, GenCo outperforms previous state-of-the-art methods on multiple benchmark datasets, even when only limited in-domain text data is available.

Details

Language :
English
Database :
OpenAIRE
Accession number :
edsair.doi.dedup.....4d17b437a603fd4a7dcaca63f831ecdf