Back to Search
Start Over
Kernel discriminant analysis for positive definite and indefinite kernels
- Source :
- IEEE Transactions on Pattern Analysis and Machine Intelligence. June, 2009, Vol. 31 Issue 6, p1017, 15 p.
- Publication Year :
- 2009
-
Abstract
- Kernel methods are a class of well established and successful algorithms for pattern analysis due to their mathematical elegance and good performance. Numerous nonlinear extensions of pattern recognition techniques have been proposed so far based on the so-called kernel trick. The objective of this paper is twofold. First, we derive an additional kernel tool that is still missing, namely kernel quadratic discriminant (KQD). We discuss different formulations of KQD based on the regularized kernel Mahalanobis distance in both complete and class-related subspaces. Second, we propose suitable extensions of kernel linear and quadratic discriminants to indefinite kernels. We provide classifiers that are applicable to kernels defined by any symmetric similarity measure. This is important in practice because problem-suited proximity measures often violate the requirement of positive definiteness. As in the traditional case, KQD can be advantageous for data with unequal class spreads in the kernel-induced spaces, which cannot be well separated by a linear discriminant. We illustrate this on artificial and real data for both positive definite and indefinite kernels. Index Terms--Machine learning, pattern recognition, kernel methods, indefinite kernels, discriminant analysis.
Details
- Language :
- English
- ISSN :
- 01628828
- Volume :
- 31
- Issue :
- 6
- Database :
- Gale General OneFile
- Journal :
- IEEE Transactions on Pattern Analysis and Machine Intelligence
- Publication Type :
- Academic Journal
- Accession number :
- edsgcl.200720214