Abstract | ||
---|---|---|
The inverse of normal covariance matrix is called precision matrix and often plays an important role in statistical estimation problem. This paper deals with the problem of estimating the precision matrix under a quadratic loss, where the precision matrix is restricted to a bounded parameter space. Gauss’ divergence theorem with matrix argument shows that the unbiased and unrestricted estimator is dominated by a posterior mean associated with a flat prior on the bounded parameter space. Also, an improving method is given by considering an expansion estimator. A hierarchical prior is shown to improve on the posterior mean. An application is given for a Bayesian prediction in a random-effects model. |
Year | DOI | Venue |
---|---|---|
2014 | 10.1016/j.jmva.2014.02.016 | Journal of Multivariate Analysis |
Keywords | Field | DocType |
primary,secondary | Econometrics,Estimation of covariance matrices,Matrix (mathematics),Bayesian hierarchical modeling,Matrix gamma distribution,Covariance matrix,Statistics,Bayes estimator,Mathematics,Bounded function,Estimator | Journal |
Volume | ISSN | Citations |
127 | 0047-259X | 0 |
PageRank | References | Authors |
0.34 | 0 | 1 |
Name | Order | Citations | PageRank |
---|---|---|---|
Hisayuki Tsukuma | 1 | 8 | 4.66 |