Distributions

Geometric Distribution

The geometric distribution frequently occurs in applications because we are often interested in how long we have to wait before a certain event happens.

A random variable X for which

P[X=i]=(1p)i1p, for i=1,2,3,

is said to have the geometric distribution with parameter p, abbreviated as XGeometric(p).

As a santiy check, we can verify the total probability of X is equal to 1:

i=1P[X=i]=i=1(1p)i1p=pi=1(1p)i1=p×11(1p)=1

If we plot the distribution of X, we get a curve that decreases monotonically by a factor of 1p at each step.

Mean and Variance

Let X be a random variable that takes in values {0,1,2,}. Then,

E[X]=i=1P[Xi]

Below is the proof of what we call the Tail Sum Formula.

Proof

For notational convenience, let’s write pi=P[X=i], for i=0,1,2,.

E[X]=(0×p0)+(1×p1)+(2×p2)+=p1+(p2+p2)+(p3+p3+p3)+=(p1+p2+p3+)+(p2+p3+)+=P[X1]+P[X2]+P[X3]+

In compact notation:

E[X]=j=1j×P[X=j]=j=1ji=1P[X=j]=i=1j=iPPX=j]=i=1P[X1]

For XGeometric(p), we have E[X]=1p.

Proof

The key observation is for a geometric random variable X,

P[Xi]=(1p)i1, for i=1,2,3,

We can obtain this simply by summing P[X=j] for ji. Another way of seeing this is to note that the event Xi means at least i tosses are required, or the first i1 tosses are all tails and the probability of this event is (1p)i1.

E[X]=i=1P[Xi]=i=1(1p)i1=11(1p)=1p

And for XGeometric(p), we have Var(X)=1pp2.

Proof

We will show E[X(X1)]=2(1p)p2:

Var(X)=E[X2]E[X]2=E[X(X1)]+E[X]E[X]2=2(1p)p2+1p1p2=2(1p)+p1p2=1pp2

To show E[X(X1)]=2(1p)p2:

d2xdy2(i=0(1p)i=d2xdy2(1p)=i=2i(i1)(1p)i2=2p3
E[X(X1)]=i=1i(i1)×P[X=i]=i=2i(i1)(1p)i1p=p(1p)i=2i(i1)(1p)i2=p(1p)×2p3=2(1p)p2

Poisson Distribution

A random variable X for which

P[X=i]=λii!eλ, for i=0,1,2,

is said to have the Poisson distribution with parameter λ, abbreviated as XPoisson(λ).

As a sanity check,

i=0λii!eλ=eλi=0λii!=eλ×eλ=1

where in the second to last step, we used the Taylor expansion ex=1+x+x22!+.

The Poisson distribution is used as a widely accepted model for so-called “rare events”, such as car accidents or ICU admission at a hospital. It is appropriate whenever the occurrences can be assumed to happen randomly with some constant density in a continuous region (of time or space), such that occurrences in disjoint subregions are independent.

Mean and Variance

For a Poisson random variable X with parameter λ, we have E[X]=Var(X)=λ,

Proof

E[X]=i=0i×P[X=i]=i=1i×λii!eλ=λeλi=iλi1(i1)!=λeλeλ=λ

Similarly,

E[X(X1)]=i=0i(i1)×[P][X=i]=i=2i(i1)λii!eλ=λ2eλi=2λi2(i2)!=λ2eλeλ=λ2

Therefore,

Var(X)=E[X2]E[X]2=E[X(X1)]+E[X]E[X]2=λ2+λλ2=λ

Sum of Independent Poisson Random Variables

Theorem

Let XPoisson(λ) and YPoisson(μ) be independent Poisson random variables. Then, X+YPoisson(λ+μ).

Proof

For all k=0,1,2,, we have

P[X+Y=k]=kj=0P[X=j,Y=kj]=kj=0P[X=j]P[Y=kj]=kj=0λjj!eλμkj(kj)!eμ=e(λ+μ)1k!kj=0k!j!(kj)!λjμkj=e(λ+μ)(λ+μ)kk!

where the second equality follows from independence, and the last equality from the binomial theorem.

By induction, we can conclude for X1,X2,,Xn independent Poisson random variables with parameters λ1,,λn.,

X1++XnPoisson(λ1++λn)

Poisson as a Limit of Binomial

Theorem

Let XBinomial(n,λn) where λ>0 is a fixed constant. For every i=0,1,2,,

PX=iλii!eλ as n

Proof

Fix i{0,1,2,}, and assume ni (because n). Then, because X is a binomial distribution,

\mathbb{P}[X=i]=\binom{n}{i}p^i(1-p)^{n-i}=\frac{n!}{i!(n-i)!}(\frac{\lambda} {n})^i(1-\frac{\lambda}{n})^{n-i}

Collecting the factors,

\mathbb{P}[X=i]=\frac{\lambda^i}{i!}(\frac{n!}{(n-i)!}\frac{1}{n^i})(1-\frac {\lambda}{n})^n(1-\frac{\lambda}{n})^{-i}

The first parenthesis above, as n\rightarrow\infty, becomes

\frac{n!}{(n-i)!}\frac{1}{n^i}=\frac{n(n-1)\cdots(n-i+1)(n-i)!}{(n-i)!}\frac {1}{n^i}=\frac{n}{n}\frac{n-1}{n}\cdots\frac{n-i+1}{n}\rightarrow 1

The second partnhesis becomes

(1-\frac{\lambda}{n})^n\rightarrow e^{-\lambda}

Since i is fixed,

(1-\frac{\lambda}{n})^{-i}\rightarrow(1-0)^{-i}=1

Substituting these into the above equation,

\mathbb{P}[X=i]\rightarrow\frac{\lambda^i}{i!}e^-{\lambda}