G-prior

From testwiki
Revision as of 19:48, 26 July 2024 by imported>Bender235 (Further reading)
(diff) ← Older revision | Latest revision (diff) | Newer revision → (diff)
Jump to navigation Jump to search

Template:Short description Template:Lowercase title In statistics, the g-prior is an objective prior for the regression coefficients of a multiple regression. It was introduced by Arnold Zellner.[1] It is a key tool in Bayes and empirical Bayes variable selection.[2][3]

Definition

Consider a data set (x1,y1),,(xn,yn), where the xi are Euclidean vectors and the yi are scalars. The multiple regression model is formulated as

yi=xiβ+εi.

where the εi are random errors. Zellner's g-prior for β is a multivariate normal distribution with covariance matrix proportional to the inverse Fisher information matrix for β, similar to a Jeffreys prior.

Assume the εi are i.i.d. normal with zero mean and variance ψ1. Let X be the matrix with ith row equal to xi. Then the g-prior for β is the multivariate normal distribution with prior mean a hyperparameter β0 and covariance matrix proportional to ψ1(XX)1, i.e.,

β|ψN[β0,gψ1(XX)1].

where g is a positive scalar parameter.

Posterior distribution of beta

The posterior distribution of β is given as

β|ψ,x,yN[qβ^+(1q)β0,qψ(XX)1].

where q=g/(1+g) and

β^=(XX)1Xy.

is the maximum likelihood (least squares) estimator of β. The vector of regression coefficients β can be estimated by its posterior mean under the g-prior, i.e., as the weighted average of the maximum likelihood estimator and β0,

β~=qβ^+(1q)β0.

Clearly, as g →∞, the posterior mean converges to the maximum likelihood estimator.

Selection of g

Estimation of g is slightly less straightforward than estimation of β. A variety of methods have been proposed, including Bayes and empirical Bayes estimators.[3]

References

Template:Reflist

Further reading