Title
A Smoothing Proximal Gradient Algorithm for Nonsmooth Convex Regression with Cardinality Penalty.
Abstract
In this paper, we focus on the constrained sparse regression problem, where the loss function is convex but nonsmooth and the penalty term is defined by the cardinality function. First, we give an exact continuous relaxation problem in the sense that both problems have the same optimal solution set. Moreover, we show that a vector is a local minimizer with the lower bound property of the original problem if and only if it is a lifted stationary point of the relaxation problem. Second, we propose a smoothing proximal gradient (SPG) algorithm for finding a lifted stationary point of the continuous relaxation model. Our algorithm is a novel combination of the classical proximal gradient algorithm and the smoothing method. We prove that the proposed SPG algorithm globally converges to a lifted stationary point of the relaxation problem, has the local convergence rate of o(k(-tau)) with tau is an element of (0, 1/2) on the objective function value, and identifies the zero entries of the lifted stationary point in finite iterations. Finally, we use three examples to illustrate the validity of the continuous relaxation model and good numerical performance of the SPG algorithm.
Year
DOI
Venue
2020
10.1137/18M1186009
SIAM JOURNAL ON NUMERICAL ANALYSIS
Keywords
Field
DocType
nonsmooth convex regression,cardinality penalty,proximal gradient method,smoothing method,global sequence convergence
Mathematical optimization,Regression,Cardinality,Regular polygon,Smoothing,Mathematics
Journal
Volume
Issue
ISSN
58
1
0036-1429
Citations 
PageRank 
References 
2
0.37
0
Authors
2
Name
Order
Citations
PageRank
Wei Bian128614.65
Xiaojun Chen21298107.51