Affordable Access

M-Power Regularized Least Squares Regression

Authors
Type
Preprint
Publication Date
Submission Date
Identifiers
arXiv ID: 1310.2451
Source
arXiv
License
Yellow
External links

Abstract

Regularization is used to find a solution that both fits the data and is sufficiently smooth, and thereby is very effective for designing and refining learning algorithms. But the influence of its exponent remains poorly understood. In particular, it is unclear how the exponent of the reproducing kernel Hilbert space (RKHS) regularization term affects the accuracy and the efficiency of kernel-based learning algorithms. Here we consider regularized least squares regression (RLSR) with an RKHS regularization raised to the power of m, where m is a variable real exponent. We design an efficient algorithm for solving the associated minimization problem, we provide a theoretical analysis of its stability, and we {compare it %/ demonstrate its advantage with respect to computational complexity, speed of convergence and prediction accuracy to %/over} the classical kernel ridge regression algorithm where the regularization exponent m is fixed at 2. Our results show that the m-power RLSR problem can be solved efficiently, and support the suggestion that one can use a regularization term that grows significantly slower than the standard quadratic growth in the RKHS norm.}

Statistics

Seen <100 times