Slutsky's theorem

From testwiki
Jump to navigation Jump to search

Template:Short description In probability theory, Slutsky's theorem extends some properties of algebraic operations on convergent sequences of real numbers to sequences of random variables.[1]

The theorem was named after Eugen Slutsky.[2] Slutsky's theorem is also attributed to Harald Cramér.[3]

Statement

Let Xn,Yn be sequences of scalar/vector/matrix random elements. If Xn converges in distribution to a random element X and Yn converges in probability to a constant c, then

  • Xn+Yn d X+c;
  • XnYn d Xc;
  • Xn/Yn d X/c,   provided that c is invertible,

where d denotes convergence in distribution.

Notes:

  1. The requirement that Yn converges to a constant is important — if it were to converge to a non-degenerate random variable, the theorem would be no longer valid. For example, let XnUniform(0,1) and Yn=Xn. The sum Xn+Yn=0 for all values of n. Moreover, YndUniform(1,0), but Xn+Yn does not converge in distribution to X+Y, where XUniform(0,1), YUniform(1,0), and X and Y are independent.[4]
  2. The theorem remains valid if we replace all convergences in distribution with convergences in probability.

Proof

This theorem follows from the fact that if Xn converges in distribution to X and Yn converges in probability to a constant c, then the joint vector (Xn, Yn) converges in distribution to (Xc) (see here).

Next we apply the continuous mapping theorem, recognizing the functions g(x,y) = x + y, g(x,y) = xy, and g(x,y) = x y−1 are continuous (for the last function to be continuous, y has to be invertible).

See also

References

Template:Reflist

Further reading

  1. Template:Cite book
  2. Template:Cite journal
  3. Slutsky's theorem is also called Cramér's theorem according to Remark 11.1 (page 249) of Template:Cite book
  4. See Template:Cite web