Neyman Type A distribution

From testwiki
Jump to navigation Jump to search

Template:Short description Template:Expert needed Template:Probability distribution

In statistics and probability, the Neyman Type A distribution is a discrete probability distribution from the family of Compound Poisson distribution. First of all, to easily understand this distribution we will demonstrate it with the following example explained in Univariate Discret Distributions;[1] we have a statistical model of the distribution of larvae in a unit area of field (in a unit of habitat) by assuming that the variation in the number of clusters of eggs per unit area (per unit of habitat) could be represented by a Poisson distribution with parameter λ, while the number of larvae developing per cluster of eggs are assumed to have independent Poisson distribution all with the same parameter ϕ. If we want to know how many larvae there are, we define a random variable Y as the sum of the number of larvae hatched in each group (given j groups). Therefore, Y = X1 + X2 + ... X j, where X1,...,Xj are independent Poisson variables with parameter λ and ϕ.

History

Jerzy Neyman was born in Russia in April 16 of 1894, he was a Polish statistician who spent the first part of his career in Europe. In 1939 he developed the Neyman Type A distribution [1] to describe the distribution of larvae in experimental field plots. Above all, it is used to describe populations based on contagion, e.g., entomology (Beall[1940],[2] Evans[1953][3]), accidents (Creswell i Froggatt [1963]),[4] and bacteriology.

The original derivation of this distribution was on the basis of a biological model and, presumably, it was expected that a good fit to the data would justify the hypothesized model. However, it is now known that it is possible to derive this distribution from different models (William Feller[1943]),[5] and in view of this, Neyman's distribution derive as Compound Poisson distribution. This interpretation makes them suitable for modelling heterogeneous populations and renders them examples of apparent contagion.

Despite this, the difficulties in dealing with Neyman's Type A arise from the fact that its expressions for probabilities are highly complex. Even estimations of parameters through efficient methods, such as maximum likelihood, are tedious and not easy to understand equations.

Definition

Probability generating function

The probability generating function (pgf) G1(z), which creates N independent Xj random variables, is used to a branching process. Each Xj produces a random number of individuals, where X1, X2,... have the same distribution as X, which is that of X with pgf G2(z). The total number of  individuals is then the random variable,[1]

Y=SN=X1+X2+...+XN

The p.g.f. of the distribution of SN is :

E[zSN]=EN[E[zSN|N]]=EN[G2(z)]=G1(G2(z))

One of the notations, which is particularly helpful, allows us to use a symbolic representation to refer to  an F1 distribution that has been generalized by an F2 distribution is,

YF1FN

In this instance, it is written as,

YPois(λ)Pois(ϕ)

Finally, the probability generating function is,

GY(z)=exp(λ(eϕ(z1)1))

From the generating function of probabilities we can calculate the probability mass function explained below.

Probability mass function

Let X1,X2,...Xj be Poisson independent variables. The probability distribution of the random variable Y = X1 +X2+...Xj is the Neyman's Type A distribution with parameters λ and ϕ.

px=P(Y=x)=eλϕxx!j=0(λeϕ)jjxj!x=1,2,...

Alternatively,

px=P(Y=x)=eλ+λeϕϕxx!j=0xS(x,j)λjeϕj

In order to see how the previous expression develops, we must bear in mind that the probability mass function is calculated from the probability generating function, and use the property of Stirling Numbers. Let's see the development

G(z)=eλ+λeϕj=0λjejϕ(eϕz1)jj|
=eλ+λeϕj=0(λeϕ)jx=jS(x,j)ϕxzxx!
=eλ+λeϕϕxx!j=0xS(x,j)λjeϕj

Another form to estimate the probabilities is with recurring successions,[6]

px=P(Y=x)=λϕeϕxr=0x1ϕrr!pxr1,p0=exp(λ+λeϕ)

Although its length varies directly with n, this recurrence relation is only employed for numerical computation and is particularly useful for computer applications.

where

  • x = 0, 1, 2, ... , except for probabilities of recurring successions, where x = 1, 2, 3, ...
  • jx
  • λ, ϕ>0.
  • x! and j! are the factorials of x and j, respectively.
  • one of the properties of Stirling numbers of the second kind is as follows:[7]
(eϕz1)j=j!x=jS(x,j)ϕzxx!

Notation

Y NA(λ,ϕ)

Properties

Moment and cumulant generating functions

The moment generating function of a random variable X is defined as the expected value of et, as a function of the real parameter t. For an NA(λ,ϕ), the moment generating function exists and is equal to

M(t)=GY(et)=exp(λ(eϕ(et1)1))

The cumulant generating function is the logarithm of the moment generating function and is equal to [1]

K(t)=log(M(t))=λ(eϕ(et1)1)

In the following table we can see the moments of the order from 1 to 4

Order Moment Cumulant
1 μ1=λϕ μ
2 μ2=λϕ(1+ϕ) σ2
3 μ3=λϕ(1+3ϕ+ϕ2) k3
4 μ4=λϕ(1+7ϕ+6ϕ2+ϕ3)+3λ2ϕ2(1+ϕ)2 k4+3σ4

Skewness

The skewness is the third moment centered around the mean divided by the 3/2 power of the standard deviation, and for the NA distribution is,

γ1=μ3μ23/2=λϕ(1+3ϕ+ϕ2)(λϕ(1+ϕ))3/2

Kurtosis

The kurtosis is the fourth moment centered around the mean, divided by the square of the variance, and for the NA distribution is,

β2=μ4μ22=λϕ(1+7ϕ+6ϕ2+ϕ3)+3λ2ϕ2(1+ϕ)2λ2ϕ2(1+ϕ)2=1+7ϕ+6ϕ2+ϕ3ϕλ(1+ϕ)2+3

The excess kurtosis is just a correction to make the kurtosis of the normal distribution equal to zero, and it is the following,

γ2=μ4μ223=1+7ϕ+6ϕ2+ϕ3ϕλ(1+ϕ)2
  • Always β2>3, or γ2>0 the distribution has a high acute peak around the mean and fatter tails.

Characteristic function

In a discrete distribution the characteristic function of any real-valued random variable is defined as the expected value of eitX, where i is the imaginary unit and t ∈ R

ϕ(t)=E[eitX]=j=0eijtP[X=j]

This function is related to the moment generating function via ϕx(t)=MX(it). Hence for this distribution the characteristic function is,

ϕx(t)=exp(λ(eϕ(eit1)1))
  • Note that the symbol ϕx is used to represent the characteristic function.

Cumulative distribution function

The cumulative distribution function is,

F(x;λ,ϕ)=P(Yx)=eλi=0xϕii!j=0(λeϕ)jjij!=eλi=0xj=0ϕi(λeϕ)jjii!j!

Other properties

d=σ2μ=1+ϕ
  • From a sample of size N, where each random variable Yi comes from a NA(λ,ϕ), where Y1, Y2, .., Yn are independent. This gives the MLE estimator as,[9]
i=1NxiN=x¯=λ¯ϕ¯ where μ is the poblational mean of x¯
  • Between the two earlier expressions We are able to parametrize using μ and d,
{μ=λϕd=1+ϕλ=μd1ϕ=d1

Parameter estimation

Method of moments

The mean and the variance of the NA(λ,ϕ) are μ=λϕ and λϕ(1+ϕ), respectively. So we have these two equations,[10]

{x¯=λϕs2=λϕ(1+ϕ)
  • s2 and x¯ are the mostral variance and mean respectively.

Solving these two equations we get the moment estimators λ^ and ϕ^ of λ and ϕ.

λ¯=x¯2s2x¯
ϕ¯=s2x¯x¯

Maximum likelihood

Calculating the maximum likelihood estimator of λ and ϕ involves multiplying all the probabilities in the probability mass function to obtain the expression (λ,ϕ;x1,,xn).

When we apply the parameterization adjustment defined in "Other Properties," we get (μ,d;X). We may define the Maximum likelihood estimation based on a single parameter if we estimate the μ as the x¯ (sample mean) given a sample X of size N. We can see it below.

(d;X)=i=1nP(xi;d)
  • To estimate the probabilities, we will use the p.m.f. of recurring successions, so that the calculation is less complex.

Testing Poisson assumption

When NA(λ,ϕ) is used to simulate a data sample it is important to see if the Poisson distribution fits the data well. For this, the following Hypothesis test is used:

{H0:d=1H1:d>1

Likelihood-ratio test

The likelihood-ratio test statistic for NA is,

W=2((X;μ,d)(X;μ,1))

Where likelihood () is the log-likelihood function. W does not have an asymptotic χ12 distribution as expected under the null hypothesis since d = 1 is at the parameter domain's edge. In the asymptotic distribution of W, it can be demonstrated that the constant 0 and χ12 have a 50:50 mixture. For this mixture, the α upper-tail percentage points are the same as the 2α upper-tail percentage points for a χ12

The poisson distribution (on { 0, 1, 2, 3, ... }) is a special case of the Neyman Type A distribution, with

Pois(λ)=NA(λ,0).

From moments of order 1 and 2 we can write the population mean and variance based on the parameters λ and ϕ.

μ=λϕ
σ2=λϕ(1+ϕ)

In the dispersion index d we observe that by substituting μ for the parametrized equation of order 1 and σ for the one of order 2, we obtain d=1+ϕ. Our variable Y is therefore distributed as a Poisson of parameter λ when d approaches to 1.

Then we have that,

limd1NA(λ,ϕ)Pois(λ)

Applications

Usage History

When the Neyman's type A species' reproduction results in clusters, a distribution has been used to characterize the dispersion of plants. This typically occurs when a species develops from offspring of parent plants or from seeds that fall close to the parent plant. However, Archibald (1948][11] observed that there is insufficient data to infer the kind of reproduction from the type of fitted distribution. While Neyman type A produced positive results for plant distributions, Evans (1953][3] showed that Negative binomial distribution produced positive results for insect distributions. Neyman type A distributions have also been studied in the context of ecology, and the results that unless the plant clusters are so compact as to not lie across the edge of the square used to pick sample locations, the distribution is unlikely to be applicable to plant populations. The compactness of the clusters is a hidden assumption in Neyman's original derivation of the distribution, according to Skellam (1958).[12] The results were shown to be significantly impacted by the square size selection.

In the context of bus driver accidents, Cresswell and Froggatt (1963) [4] derived the Neyman type A based on the following hypotheses:

  1. Each driver is susceptible to "spells," the number of which is Poissonian for any given amount of time, with the same parameter λ for all drivers.
  2. A driver's performance during a spell is poor, and he is likely to experience a Poissonian number of collisions, with the same parameter ϕ for all drivers.
  3. Each driver behaves independently.
  4. No accidents can occur outside of a spell.

These assumptions lead to a Neyman type A distribution via the Pois(λ) and Pois(ϕ) model. In contrast to their "short distribution," Cresswell and Froggatt called this one the "long distribution" because of its lengthy tail. According to Irwin(1964), a type A distribution can also be obtained by assuming that various drivers have different levels of proneness, or K. with probability:

P(K=kϕ)=eλλkk!
  • taking values 0,ϕ,2ϕ,3ϕ,...

and that a driver with proneness kϕ has X accidents where:

P(X=x|kϕ)=ekϕ(kϕ)xx!

This is the Pois(kϕ)Pois(λ) model with mixing over the values taken by K.

Distribution was also suggested in the application for the grouping of minority tents for food from 1965 to 1969. In this regard, it was predicted that only the clustering rates or the average number of entities per grouping needed to be approximated, rather than adjusting distributions on d very big data bases.

  • The reference of the usage history.[1]

Calculating Neyman Type A probabilities in R

  • The code below simulates 5000 instances of NA(λ,ϕ),
 rNeymanA <- function(n,lambda, phi){
   r <- numeric()
   for (j in 1:n) {
      k = rpois(1,lambda)
      r[j] <- sum(rpois(k,phi)) 
   }
   return(r)
 }
  • The mass function of recurring probabilities is implemented in R in order to estimate theoretical probabilities; we can see it below,
dNeyman.rec <- function(x, lambda, phi){
 p <- numeric()
 p[1]<- exp(-lambda + lambda*exp(-phi))
 c <- lambda*phi*exp(-phi)
 if(x == 0){
   return(p[1])
 }
 else{
   for (i in 1:x) {
     suma = 0
     for (r in 0:(i-1)) {
       suma = suma + (phi^r)/(factorial(r))*p[i-r]
     }
     p[i+1] = (c/(i))*suma # +1 per l'R
   }
   res <- p[i+1]
   return(res)
 }
}

We compare results between the relative frequencies obtained by the simulation and the probabilities computed by the p.m.f. . Given two values for the parameters λ=2 and ϕ=1. It is displayed in the following table,

X 0 1 2 3 4 5 6 7 8
Estimated fi .271 .207 .187 .141 .082 .052 .028 .016 .008
Teoric P(X=xi) .282 .207 .180 .129 .084 .051 .029 .016 .008

References

Template:Reflist Template:ProbDistributions