Back to Search
Start Over
Solving the Imbalanced Problem by Metric Learning and Oversampling
- Source :
- IEEE Transactions on Knowledge and Data Engineering; December 2024, Vol. 36 Issue: 12 p9294-9307, 14p
- Publication Year :
- 2024
-
Abstract
- Imbalanced data poses a substantial challenge to conventional classification methods, which often disproportionately favor samples from the majority class. To mitigate this issue, various oversampling techniques have been deployed, but opportunities for optimizing data distributions remain underexplored. By exploiting the ability of metric learning to refine the sample distribution, we propose a novel approach, Imbalance Large Margin Nearest Neighbor (ILMNN). Initially, ILMNN is applied to establish a latent feature space, pulling intra-class samples closer and distancing inter-class samples, thereby amplifying the efficacy of oversampling techniques. Subsequently, we allocate varying weights to samples contingent upon their local distribution and relative class frequency, thereby equalizing contributions from minority and majority class samples. Lastly, we employ Kullback-Leibler (KL) divergence as a safeguard to maintain distributional similarity to the original dataset, mitigating severe intra-class imbalances. Comparative experiments on various class-imbalanced datasets verify that our ILMNN approach yields superior results.
Details
- Language :
- English
- ISSN :
- 10414347 and 15582191
- Volume :
- 36
- Issue :
- 12
- Database :
- Supplemental Index
- Journal :
- IEEE Transactions on Knowledge and Data Engineering
- Publication Type :
- Periodical
- Accession number :
- ejs67986145
- Full Text :
- https://doi.org/10.1109/TKDE.2024.3419834