Back to Search Start Over

Filter-based unsupervised feature selection using Hilbert–Schmidt independence criterion

Authors :
Samaneh Liaghat
Eghbal G. Mansoori
Source :
International Journal of Machine Learning and Cybernetics. 10:2313-2328
Publication Year :
2018
Publisher :
Springer Science and Business Media LLC, 2018.

Abstract

Feature selection is a fundamental preprocess before performing actual learning; especially in unsupervised manner where the data are unlabeled. Essentially, when there are too many features in the problem, dimensionality reduction through discarding weak features is highly desirable. In this paper, we present a framework for unsupervised feature selection based on dependency maximization between the samples similarity matrices before and after deleting a feature. In this regard, a novel estimation of Hilbert–Schmidt independence criterion (HSIC), more appropriate for high-dimensional data with small sample size, is introduced. Its key idea is that by eliminating the redundant features and/or those have high inter-relevancy, the pairwise samples similarity is not affected seriously. Also, to handle the diagonally dominant matrices, a heuristic trick is used in order to reduce the dynamic range of matrix values. In order to speed up the proposed scheme, the gap statistic and k-means clustering methods are also employed. To assess the performance of our method, some experiments on benchmark datasets are conducted. The obtained results confirm the efficiency of our unsupervised feature selection scheme.

Details

ISSN :
1868808X and 18688071
Volume :
10
Database :
OpenAIRE
Journal :
International Journal of Machine Learning and Cybernetics
Accession number :
edsair.doi...........69be0af1d93d3eb6ba507ec20171ed32