A wide class of regularization problems in machine learning and statistics employ a regularization term which is obtained by composing a simple convex function omega with a linear transformation. This setting includes Group Lasso methods, the Fused Lasso and other total variation methods, multi-task learning methods and many more. In this paper, we present a general approach for computing the proximity operator of this class of regularizers, under the assumption that the proximity operator of the function \omega is known in advance. Our approach builds on a recent line of research on optimal first order optimization methods and uses fixed point iterations for numerically computing the proximity operator. It is more general than current approaches and, as we show with numerical simulations, computationally more efficient than available first order methods which do not achieve the optimal rate. In particular, our method outperforms state of the art O(1/T) methods for overlapping Group Lasso and matches optimal O(1/T2) methods for the Fused Lasso and tree structured Group Lasso.
Argyriou, Andreas; Micchelli, Charles A.; Pontil, Massimiliano; Shen, Lixin; and Xu, Yuesheng, "Efficient First Order Methods for Linear Composite Regularizers" (2011). Mathematics Faculty Scholarship. Paper 71.
Harvested from arXiv.org
Creative Commons License
This work is licensed under a Creative Commons Attribution 3.0 License.