Back to Search Start Over

Leveraging Biases in Large Language Models: 'bias-kNN' for Effective Few-Shot Learning

Authors :
Zhang, Yong
Li, Hanzhang
Li, Zhitao
Cheng, Ning
Li, Ming
Xiao, Jing
Wang, Jianzong
Zhang, Yong
Li, Hanzhang
Li, Zhitao
Cheng, Ning
Li, Ming
Xiao, Jing
Wang, Jianzong
Publication Year :
2024

Abstract

Large Language Models (LLMs) have shown significant promise in various applications, including zero-shot and few-shot learning. However, their performance can be hampered by inherent biases. Instead of traditionally sought methods that aim to minimize or correct these biases, this study introduces a novel methodology named ``bias-kNN''. This approach capitalizes on the biased outputs, harnessing them as primary features for kNN and supplementing with gold labels. Our comprehensive evaluations, spanning diverse domain text classification datasets and different GPT-2 model sizes, indicate the adaptability and efficacy of the ``bias-kNN'' method. Remarkably, this approach not only outperforms conventional in-context learning in few-shot scenarios but also demonstrates robustness across a spectrum of samples, templates and verbalizers. This study, therefore, presents a unique perspective on harnessing biases, transforming them into assets for enhanced model performance.<br />Comment: Accepted by the 49th IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP 2024)

Details

Database :
OAIster
Publication Type :
Electronic Resource
Accession number :
edsoai.on1438516624
Document Type :
Electronic Resource