Title | ||
---|---|---|
Approximate Message Passing With Consistent Parameter Estimation and Applications to Sparse Learning |
Abstract | ||
---|---|---|
We consider the estimation of an independent and identically distributed (i.i.d.) (possibly non-Gaussian) vector x ∈ Rn from measurements y ∈ Rm obtained by a general cascade model consisting of a known linear transform followed by a probabilistic componentwise (possibly nonlinear) measurement channel. A novel method, called adaptive generalized approximate message passing (adaptive GAMP) is presented. It enables the joint learning of the statistics of the prior and measurement channel along with estimation of the unknown vector x. We prove that, for large i.i.d. Gaussian transform matrices, the asymptotic componentwise behavior of the adaptive GAMP is predicted by a simple set of scalar state evolution equations. In addition, we show that the adaptive GAMP yields asymptotically consistent parameter estimates, when a certain maximum-likelihood estimation can be performed in each step. This implies that the algorithm achieves a reconstruction quality equivalent to the oracle algorithm that knows the correct parameter values. Remarkably, this result applies to essentially arbitrary parametrizations of the unknown distributions, including nonlinear and non-Gaussian ones. The adaptive GAMP methodology thus provides a systematic, general and computationally efficient method applicable to a large range of linear-nonlinear models with provable guarantees. |
Year | DOI | Venue |
---|---|---|
2012 | 10.1109/TIT.2014.2309005 | IEEE Transactions on Information Theory |
Keywords | Field | DocType |
gaussian processes,compressed sensing,maximum likelihood estimation,message passing,adaptive gamp,adaptive generalized approximate message passing,compressive sensing,i.i.d. gaussian transform matrices,independent and identically distributed vector,linear transform,maximum-likelihood estimation,nongaussian vector,oracle algorithm,parameter estimation,probabilistic componentwise measurement channel,scalar state evolution equations,sparse learning,approximate message passing,belief propagation,mathematical model,approximation algorithms,vectors,estimation | Mathematical optimization,Nonlinear system,Computer science,Matrix (mathematics),Gaussian,Gaussian process,Independent and identically distributed random variables,Artificial intelligence,Estimation theory,Machine learning,Message passing,Belief propagation | Journal |
Volume | Issue | ISSN |
60 | 5 | 0018-9448 |
Citations | PageRank | References |
36 | 1.18 | 27 |
Authors | ||
4 |
Name | Order | Citations | PageRank |
---|---|---|---|
Ulugbek Kamilov | 1 | 89 | 6.77 |
Sundeep Rangan | 2 | 3101 | 163.90 |
Alyson K. Fletcher | 3 | 552 | 41.10 |
M Unser | 4 | 4335 | 499.89 |