Difference between revisions of "Poisson distribution"
From Maths
m |
m |
||
Line 23: | Line 23: | ||
*** the first 2 terms are easy to give: {{M|e^{\lambda} }} and {{M|\lambda e^{-\lambda} }} respectively, after that we have {{M|\frac{1}{2}\lambda^2 e^{-\lambda} }} and so forth | *** the first 2 terms are easy to give: {{M|e^{\lambda} }} and {{M|\lambda e^{-\lambda} }} respectively, after that we have {{M|\frac{1}{2}\lambda^2 e^{-\lambda} }} and so forth | ||
** for {{M|k\in\mathbb{N}_{\ge 0} }} we have: {{MM|\mathbb{P}[X\le k]\eq e^{-\lambda}\sum^k_{j\eq 0}\frac{1}{j!}\lambda^j}} | ** for {{M|k\in\mathbb{N}_{\ge 0} }} we have: {{MM|\mathbb{P}[X\le k]\eq e^{-\lambda}\sum^k_{j\eq 0}\frac{1}{j!}\lambda^j}} | ||
− | + | ===[[/RV|As a formal random variable]]=== | |
+ | {{/RV|home=true}} | ||
==To add to page== | ==To add to page== | ||
* [[The sum of two random variables with Poisson distributions is a Poisson distribution itself]] | * [[The sum of two random variables with Poisson distributions is a Poisson distribution itself]] |
Revision as of 15:19, 22 November 2017
Stub grade: A*
This page is a stub
This page is a stub, so it contains little or minimal information and is on a to-do list for being expanded.The message provided is:
My informal derivation feels too formal, but isn't formal enough to be a formal one! Work in progress!
Poisson distribution | |
X∼Poi(λ) | |
λ∈R≥0 (λ - the average rate of events per unit) | |
file.png | |
Definition | |
---|---|
Type | Discrete, over N≥0 |
p.m.f | P[X=k]:=e−λλkk! |
c.d.f | P[X≤k]=e−λk∑i=0λik! |
Characteristics | |
Expected value | E[X]=λ |
Variance | Var(X)=λ |
Contents
[hide]Definition
- X∼Poisson(λ)
- for k∈N≥0 we have: P[X=k]:=e−λλkk!
- the first 2 terms are easy to give: eλ and λe−λ respectively, after that we have 12λ2e−λ and so forth
- for k∈N≥0 we have: P[X≤k]=e−λk∑j=01j!λj
- for k∈N≥0 we have: P[X=k]:=e−λλkk!
As a formal random variable
- Caveat:λ here is used to denote 2 things - the parameter to the Poisson distribution, and the restriction of the 1 dimensional Lebesgue measure to some region of interest.
There is no unique way to define a random variable, here is one way.
- Let ([0,1], B([0,1]), λ) be a probability space - which itself could be viewed as a rectangular distribution's random variable
- Let λ∈R>0 be given, and let X∼Poi(λ)
- Specifically consider (N0, P(N0)) as a sigma-algebra and X:[0,1]→N0 by:
- X:x↦{0if x∈[0,p1)1if x∈[p1,p2)⋮⋮kif x∈[pk,pk+1)⋮⋮ for p1:=e−λλ11! and pk:=pk−1+e−λλkk!
- Specifically consider (N0, P(N0)) as a sigma-algebra and X:[0,1]→N0 by:
- Let λ∈R>0 be given, and let X∼Poi(λ)
Giving the setup shown on the left.
To add to page
- The sum of two random variables with Poisson distributions is a Poisson distribution itself
- If X∼Poi(λ) and Y∼Poi(λ′) then X+Y∼Poi(λ+λ′)
- Time between events of a Poisson distribution Alec (talk) 22:06, 11 November 2017 (UTC)
Mean
- ∞∑n=0n×P[X=n]=∞∑n=0[n×e−λλnn!]=0+[e−λ∞∑n=1λn(n−1)!]=e−λλ[∞∑n=1λn−1(n−1)!]
- =λe−λ[∞∑n=0λnn!]=λe−λ[limn→∞(n∑k=0λkk!)]
- But! ex=limn→∞(n∑i=0xii!)
- So =λe−λeλ
- =λ
- =λe−λ[∞∑n=0λnn!]=λe−λ[limn→∞(n∑k=0λkk!)]
Derivation
Standard Poisson distribution:
- Let S:=[0,1)⊆R, recall that means S={x∈R | 0≤x<1}
- Let λ be the average count of some event that can occur 0 or more times on S
We will now divide S up into N equally sized chunks, for N∈N≥1
- Let Si,N:=[i−1N,iN)[Note 1] for i∈{1,…,N}⊆N
We will now define a random variable that counts the occurrences of events per interval.
- Let C(Si,N) be the RV such that its value is the number of times the event occurred in the [i−1N,iN) interval
We now require:
- limN→∞(P[C(Si,N)≥2])=0 - such that:
- as the Si,N get smaller the chance of 2 or more events occurring in the space reaches zero.
- Warning:This is phrased as a limit, I'm not sure it should be as we don't have any Si,∞ so no BORV(λN) distribution then either
Note that:
- limN→∞(C(Si,N))=limN→∞(BORV(λN))
- This is supposed to convey that the distribution of C(Si,N) as N gets large gets arbitrarily close to BORV(λN)
So we may say for sufficiently large N that:
- C(Si,N)∼(approx)BORV(λN), so that:
- P[C(Si,N)=0]≈(1−λN)
- P[C(Si,N)=1]≈λN, and of course
- P[C(Si,N)≥2]≈0
Assuming the C(Si,N) are independent over i (which surely we get from the BORV distributions?) we see:
- C(S)∼(approx)Bin(N,λN) or, more specifically: C(S)=limN→∞(N∑i=1C(Si,N))=limN→∞(Bin(N,λN))
We see:
- P[C(S)=k]=limN→∞(P[Bin(N,λN)=k])=limN→∞(NCk (λN)k(1−λN)N−k)
We claim that:
- limN→∞(NCk (λN)k(1−λN)N−k)=λkk!e−λ
We will tackle this in two parts:
- limN→∞(NCk (λN)k⏟A (1−λN)N−k⏟B) where B\rightarrow e^{-\lambda} and A\rightarrow \frac{\lambda^k}{k!}
Proof
Key notes:
A
Notice:
- {}^N\!C_k\ \left(\frac{\lambda}{N}\right)^k \eq \frac{N!}{(N-k)!k!}\cdot\frac{1}{N^k}\cdot\lambda^k
- \eq\frac{1}{k!}\cdot\frac{\overbrace{N(N-1)\cdots(N-k+2)(N-k+1)}^{k\text{ terms} } }{\underbrace{N\cdot N\cdots N}_{k\text{ times} } } \cdot\lambda^k
- Notice that as N gets bigger N-k+1 is "basically" N so the Ns in the denominator cancel (in fact the value will be slightly less than 1, tending towards 1 as N\rightarrow\infty) this giving:
- \frac{\lambda^k}{k!}
B
This comes from:
- e^x:\eq\lim_{n\rightarrow\infty}\left(\left(1+\frac{x}{n}\right)^n\right), so we get the e^{-\lambda} term.
Notes
- Jump up ↑ Recall again that means \{x\in\mathbb{R}\ \vert\ \frac{i-1}{N}\le x < \frac{i}{N} \}