G-prior
Template:Short description Template:Lowercase title In statistics, the g-prior is an objective prior for the regression coefficients of a multiple regression. It was introduced by Arnold Zellner.[1] It is a key tool in Bayes and empirical Bayes variable selection.[2][3]
Definition
Consider a data set , where the are Euclidean vectors and the are scalars. The multiple regression model is formulated as
where the are random errors. Zellner's g-prior for is a multivariate normal distribution with covariance matrix proportional to the inverse Fisher information matrix for , similar to a Jeffreys prior.
Assume the are i.i.d. normal with zero mean and variance . Let be the matrix with th row equal to . Then the g-prior for is the multivariate normal distribution with prior mean a hyperparameter and covariance matrix proportional to , i.e.,
where g is a positive scalar parameter.
Posterior distribution of beta
The posterior distribution of is given as
where and
is the maximum likelihood (least squares) estimator of . The vector of regression coefficients can be estimated by its posterior mean under the g-prior, i.e., as the weighted average of the maximum likelihood estimator and ,
Clearly, as g →∞, the posterior mean converges to the maximum likelihood estimator.
Selection of g
Estimation of g is slightly less straightforward than estimation of . A variety of methods have been proposed, including Bayes and empirical Bayes estimators.[3]