151. A Computational Framework for Multivariate Convex Regression and Its Variants
- Author
-
Bodhisattva Sen, Rahul Mazumder, Arkopal Choudhury, Garud Iyengar, Sloan School of Management, and Mazumder, Rahul
- Subjects
Statistics and Probability ,FOS: Computer and information sciences ,Multivariate statistics ,Statistics::Theory ,Regression function ,MathematicsofComputing_NUMERICALANALYSIS ,01 natural sciences ,Statistics - Computation ,Methodology (stat.ME) ,Statistics::Machine Learning ,010104 statistics & probability ,Computer Science::Systems and Control ,0502 economics and business ,FOS: Mathematics ,Statistics::Methodology ,Applied mathematics ,Quadratic programming ,0101 mathematics ,Mathematics - Optimization and Control ,Computation (stat.CO) ,Statistics - Methodology ,050205 econometrics ,Mathematics ,Augmented Lagrangian method ,05 social sciences ,Regular polygon ,Nonparametric statistics ,Regression ,Optimization and Control (math.OC) ,Statistics, Probability and Uncertainty - Abstract
We study the nonparametric least squares estimator (LSE) of a multivariate convex regression function. The LSE, given as the solution to a quadratic program with O(n²) linear constraints (n being the sample size), is difficult to compute for large problems. Exploiting problem specific structure, we propose a scalable algorithmic framework based on the augmented Lagrangian method to compute the LSE. We develop a novel approach to obtain smooth convex approximations to the fitted (piecewise affine) convex LSE and provide formal bounds on the quality of approximation. When the number of samples is not too large compared to the dimension of the predictor, we propose a regularization scheme—Lipschitz convex regression—where we constrain the norm of the subgradients, and study the rates of convergence of the obtained LSE. Our algorithmic framework is simple and flexible and can be easily adapted to handle variants: estimation of a nondecreasing/nonincreasing convex/concave (with or without a Lipschitz bound) function. We perform numerical studies illustrating the scalability of the proposed algorithm—on some instances our proposal leads to more than a 10,000-fold improvement in runtime when compared to off-the-shelf interior point solvers for problems with n = 500. Keywords: Augmented Lagrangian method; Lipschitz convex regression; Non parametric least squares estimator; Scalable quadratic programming; Smooth convex regression, United States. Office of Naval Research (Grant N00014-15-1-2342)
- Published
- 2019
- Full Text
- View/download PDF