Title
Linear Complementarity for Regularized Policy Evaluation and Improvement.
Abstract
Recent work in reinforcement learning has emphasized the power of L1 regularization to perform feature selection and prevent overfitting. We propose formulating the L1 regularized linear fixed point problem as a linear complementarity problem (LCP). This formulation offers several advantages over the LARS-inspired formulation, LARS-TD. The LCP formulation allows the use of efficient off-the-shelf solvers, leads to a new uniqueness result, and can be initialized with starting points from similar problems (warm starts). We demonstrate that warm starts, as well as the efficiency of LCP solvers, can speed up policy iteration. Moreover, warm starts permit a form of modified policy iteration that can be used to approximate a greedy" homotopy path, a generalization of the LARS-TD homotopy path that combines policy evaluation and optimization."
Year
Venue
Field
2010
NIPS
Complementarity (molecular biology),Uniqueness,Mathematical optimization,Feature selection,Computer science,Regularization (mathematics),Overfitting,Homotopy,Linear complementarity problem,Reinforcement learning
DocType
Citations 
PageRank 
Conference
25
1.38
References 
Authors
9
3
Name
Order
Citations
PageRank
Johns, Jeffrey1251.38
Christopher Painter-Wakefield21707.96
Ronald Parr32428186.85