De-sparsified lasso

From testwiki
Jump to navigation Jump to search

Template:Short description Template:Multiple issues

De-sparsified lasso contributes to construct confidence intervals and statistical tests for single or low-dimensional components of a large parameter vector in high-dimensional model.[1]

High-dimensional linear model

Y=Xβ0+ϵ with n×p design matrix X=:[X1,...,Xp] (n×p vectors Xj), ϵNn(0,σϵ2I) independent of X and unknown regression p×1 vector β0.

The usual method to find the parameter is by Lasso: β^n(λ)=argminβp 12nYXβ22+λβ1

The de-sparsified lasso is a method modified from the Lasso estimator which fulfills the Karush–Kuhn–Tucker conditions[2] is as follows:

β^n(λ,M)=β^n(λ)+1nMXT(YXβ^n(λ))

where MRp×p is an arbitrary matrix. The matrix M is generated using a surrogate inverse covariance matrix.

Generalized linear model

Desparsifying l1-norm penalized estimators and corresponding theory can also be applied to models with convex loss functions such as generalized linear models.

Consider the following 1×pvectors of covariables xiχRp and univariate responses yiYR for i=1,...,n

we have a loss function ρβ(y,x)=ρ(y,xβ)(βRp) which is assumed to be strictly convex function in βRp

The l1-norm regularized estimator is β^=argminβ(Pnρβ+λβ1)

Similarly, the Lasso for node wise regression with matrix input is defined as follows: Denote by Σ^ a matrix which we want to approximately invert using nodewise lasso.

The de-sparsified l1-norm regularized estimator is as follows: γj^:=argminγRp1(Σ^j,j2Σ^j,/jγ+γTΣ^/j,/jγ+2λjγ1

where Σ^j,/j denotes the jth row of Σ^ without the diagonal element (j,j), and Σ^/j,/j is the sub matrix without the jth row and jth column.

References

Template:Reflist