On Coresets for Regularized Regression

Part of Proceedings of the International Conference on Machine Learning 1 pre-proceedings (ICML 2020)

Bibtex »Metadata »Paper »Supplemental »

Bibtek download is not availble in the pre-proceeding


Authors

Rachit Chhaya, Supratim Shit, Anirban Dasgupta

Abstract

We study the effect of norm based regularization on the size of coresets for the regularized regression problems. Specifically, given a matrix $ \mathbf{A} \in {\mathbf{R}}^{n \M{x}d}$ with $n\gg d$ and a vector $\mathbf{b} \in \mathbf{R} ^ n $ and $\lambda > 0$, we analyze the size of coresets for regularized versions of regression of the form $||\M{Ax}-\M{b}||_p^r + \lambda||{\M{x}}||_q^s$ . It has been shown for the case of ridge regression ($p,q,r,s=2$) that we can obtain a coreset smaller than the coreset for its unregularized counterpart i.e. least squares regression\cite{avron2017sharper}. However we show that when $r \neq s$, no coreset for some regularized regression can have size smaller than the optimal coreset of the unregularized version. The well known LASSO problem falls under this category and hence does not allow a coreset smaller than the one for least squares regression. We propose a modified version of the LASSO problem and obtain for it a coreset of size smaller than the least square regression. We empirically show that the modified version of LASSO also induces sparsity in solution like the LASSO. We also obtain smaller coresets for $\ell_p$-regression with $\ell_p$-regularization. We extend our methods to multi response regularized regression. Finally, we empirically demonstrate the coreset performance for the modified LASSO and the $\ell_1$-regression with $\ell_1$- regularization.