Subindependence

From testwiki
Revision as of 04:43, 27 January 2023 by imported>ClueBot NG (Reverting possible vandalism by 2601:2C6:4F00:45E0:95AB:B5AC:909E:92EC to version by Monkbot. Report False Positive? Thanks, ClueBot NG. (4209630) (Bot))
(diff) ← Older revision | Latest revision (diff) | Newer revision → (diff)
Jump to navigation Jump to search

In probability theory and statistics, subindependence is a weak form of independence.

Two random variables X and Y are said to be subindependent if the characteristic function of their sum is equal to the product of their marginal characteristic functions. Symbolically:

φX+Y(t)=φX(t)φY(t).

This is a weakening of the concept of independence of random variables, i.e. if two random variables are independent then they are subindependent, but not conversely. If two random variables are subindependent, and if their covariance exists, then they are uncorrelated.[1]

Subindependence has some peculiar properties: for example, there exist random variables X and Y that are subindependent, but X and αY are not subindependent when α ≠ 1[1] and therefore X and Y are not independent.

One instance of subindependence is when a random variable X is Cauchy with location 0 and scale s and another random variable Y=X, the antithesis of independence. Then X+Y is also Cauchy but with scale 2s. The characteristic function of either X or Y in t is then exp(-s·|t|), and the characteristic function of X+Y is exp(-2s·|t|)=exp(-s·|t|)2.

Notes

Template:Reflist

References

Further reading

  1. 1.0 1.1 Hamedani & Volkmer (2009)