Back to Search Start Over

INCREMENTAL REGULARIZED LEAST SQUARES FOR DIMENSIONALITY REDUCTION OF LARGE-SCALE DATA.

Authors :
XIAOWEI ZHANG
LI CHENG
DELIN CHU
LI-ZHI LIAO
NG, MICHAEL K.
TAN, ROGER C. E.
Source :
SIAM Journal on Scientific Computing. 2016, Vol. 38 Issue 3, pB414-B439. 26p.
Publication Year :
2016

Abstract

Over the past few decades, much attention has been drawn to large-scale incremental data analysis, where researchers are faced with huge amounts of high-dimensional data acquired incrementally. In such a case, conventional algorithms that compute the result from scratch whenever a new sample comes are highly inefficient. To conquer this problem, we propose a new incremental algorithm incremental regularized least squares (IRLS) that incrementally computes the solution to the regularized least squares (RLS) problem with multiple columns on the right-hand side. More specifically, for an RLS problem with c (c > 1) columns on the right-hand side, we update its unique solution by solving an RLS problem with a single column on the right-hand side whenever a new sample arrives, instead of solving an RLS problem with c columns on the right-hand side from scratch. As a direct application of IRLS, we consider the supervised dimensionality reduction of large-scale data and focus on linear discriminant analysis (LDA). We first propose a new batch LDA model that is closely related to the RLS problem, and then apply IRLS to develop a new incremental LDA algorithm. Experimental results on real-world datasets demonstrate the effectiveness and efficiency of our algorithms. [ABSTRACT FROM AUTHOR]

Details

Language :
English
ISSN :
10648275
Volume :
38
Issue :
3
Database :
Academic Search Index
Journal :
SIAM Journal on Scientific Computing
Publication Type :
Academic Journal
Accession number :
116788697
Full Text :
https://doi.org/10.1137/15M1035653