Back to Search Start Over

Solving the Imbalanced Problem by Metric Learning and Oversampling

Authors :
Yang, Kaixiang
Yu, Zhiwen
Chen, Wuxing
Liang, Zefeng
Chen, C. L. Philip
Source :
IEEE Transactions on Knowledge and Data Engineering; December 2024, Vol. 36 Issue: 12 p9294-9307, 14p
Publication Year :
2024

Abstract

Imbalanced data poses a substantial challenge to conventional classification methods, which often disproportionately favor samples from the majority class. To mitigate this issue, various oversampling techniques have been deployed, but opportunities for optimizing data distributions remain underexplored. By exploiting the ability of metric learning to refine the sample distribution, we propose a novel approach, Imbalance Large Margin Nearest Neighbor (ILMNN). Initially, ILMNN is applied to establish a latent feature space, pulling intra-class samples closer and distancing inter-class samples, thereby amplifying the efficacy of oversampling techniques. Subsequently, we allocate varying weights to samples contingent upon their local distribution and relative class frequency, thereby equalizing contributions from minority and majority class samples. Lastly, we employ Kullback-Leibler (KL) divergence as a safeguard to maintain distributional similarity to the original dataset, mitigating severe intra-class imbalances. Comparative experiments on various class-imbalanced datasets verify that our ILMNN approach yields superior results.

Details

Language :
English
ISSN :
10414347 and 15582191
Volume :
36
Issue :
12
Database :
Supplemental Index
Journal :
IEEE Transactions on Knowledge and Data Engineering
Publication Type :
Periodical
Accession number :
ejs67986145
Full Text :
https://doi.org/10.1109/TKDE.2024.3419834