1. A rank-one update method for least squares linear discriminant analysis with concept drift
- Author
-
Yu-Chiang Frank Wang and Yi-Ren Yeh
- Subjects
Concept drift ,Rank (linear algebra) ,business.industry ,Dimensionality reduction ,Pattern recognition ,Linear discriminant analysis ,Linear subspace ,Least squares ,ComputingMethodologies_PATTERNRECOGNITION ,Artificial Intelligence ,Signal Processing ,Computer Vision and Pattern Recognition ,Artificial intelligence ,business ,Software ,Subspace topology ,Mathematics ,Curse of dimensionality - Abstract
Linear discriminant analysis (LDA) is a popular supervised dimension reduction algorithm, which projects the data into an effective low-dimensional linear subspace while the separation between the projected data from different classes is improved. While this subspace is typically determined by solving a generalized eigenvalue decomposition problem, its high computation costs prohibit the use of LDA especially when the scale and the dimensionality of the data are large. Based on the recent success of least squares LDA (LSLDA), we propose a novel rank-one update method with a simplified class indicator matrix. Using the proposed algorithm, we are able to derive the LSLDA model efficiently. Moreover, our LSLDA model can be extended to address the learning task of concept drift, in which the recently received data exhibit with gradual or abrupt changes in distribution. In other words, our LSLDA is able to observe and model the data distribution changes, while the dependency on outdated data will be suppressed. This proposed LSLDA will benefit applications of streaming data classification or mining, and it can recognize data with newly added class labels during the learning process. Experimental results on both synthetic and real datasets (with and without concept drift) confirm the effectiveness of our propose LSLDA.
- Published
- 2013