The Poisson Approximation
Why the Poisson Distribution?
The binomial distribution is the natural model for counting successes in independent trials. But when is large and is small — so that the expected count is moderate — the binomial PMF is numerically unwieldy (large factorials). The Poisson distribution provides an elegant and accurate approximation that depends on the single parameter . More than a mere computational convenience, the Poisson distribution is the natural model for rare events: the number of packet arrivals in a time slot, the number of bit errors in a long codeword, the number of active users in a massive access system.
Definition: Poisson Distribution
Poisson Distribution
A discrete random variable has the Poisson distribution with parameter , written , if:
The PMF sums to 1 because . The mean and variance are both equal to : .
Theorem: Mean and Variance of the Poisson Distribution
If , then and .
The equality of mean and variance is a distinctive fingerprint of the Poisson distribution. In data analysis, if the sample variance of count data is close to the sample mean, a Poisson model is a natural first choice.
Compute $\mathbb{E}[X]$
$
Compute $\mathbb{E}[X(X-1)]$
$
Derive the variance
$
Theorem: Poisson Limit Theorem
Let where as . Then for every fixed : That is, converges in distribution to .
When is large and is small, each trial is an almost-never event but there are many of them. The total count depends only on the product . The Poisson distribution captures this regime: it is the distribution of the total count of many independent rare events.
Write the binomial PMF
Set . Then:
Separate the factors
$
Take the limit
As with fixed:
- .
- (definition of ).
- .
Therefore:
Example: Typographical Errors
A 500-page book has 200 typographical errors scattered uniformly at random. What is the probability that a given page has no errors? Exactly 1 error?
Model
Each error independently lands on any of the 500 pages with probability . The number of errors on a given page is , with .
Poisson approximation
. .
Exact binomial (for comparison)
. The Poisson approximation agrees to four decimal places — the approximation is excellent when and .
Theorem: Le Cam's Inequality
Let be independent Bernoulli random variables with . Let and . Then: where denotes total variation distance: .
The bound says the Poisson approximation is accurate when each individual is small, even if the are not all equal. The total variation distance is at most , which is small when each . For the homogeneous case , the bound becomes , which vanishes as with fixed.
Coupling argument (sketch)
The proof constructs a coupling between and a Poisson random variable by replacing each Bernoulli with an independent Poisson . Since assigns probability to 0 and to 1, the total variation distance between and is at most .
Sum the contributions
By the triangle inequality for total variation and the additivity of independent sums under Poisson, the total variation distance between and is at most .
Le Cam for the Homogeneous Case
When for all , Le Cam's inequality gives: So if is moderate and is small (equivalently, is large), the Poisson approximation is accurate. For instance, with and (), the bound is : the total variation distance is at most 2.5%.
Poisson Approximation to the Binomial
Compare the binomial PMF to the Poisson PMF and observe convergence as grows. The total variation distance is displayed.
Parameters
Birthday Problem: Collision Probability
The birthday problem asks: in a group of people, what is the probability that at least two share a birthday? This is a Poisson-approximable rare event. Compare the exact formula with the Poisson approximation.
Parameters
Example: Prussian Horse Kicks (the Classic Example)
Ladislaus Bortkiewicz (1898) collected data on deaths from horse kicks in the Prussian army: 14 corps observed over 20 years (280 corps-years), with a total of 196 deaths. Test whether the Poisson model fits.
Estimate the rate
deaths per corps per year.
Expected frequencies
Under :
- 139 expected.
- 97.3 expected.
- 34.1 expected.
- 9.6 expected.
Compare with observed data
Observed: 144, 91, 32, 13. The Poisson model fits remarkably well — this was one of the first empirical validations of the Poisson distribution for rare events.
Example: Packet Arrivals in a Time Slot
A router receives packets from independent sources, each transmitting in a given time slot with probability . What is the probability that at most 3 packets arrive?
Compute $\lambda$
.
Poisson approximation
$
Le Cam bound
. The approximation error is at most 0.25% in total variation — excellent for engineering purposes.
Why This Matters: Poisson Model for Massive Access
In massive machine-type communication (mMTC), a large number of devices (say ) are registered, but in any given time slot each device transmits with a small probability (say ). The number of active devices is then well-modeled by with . This Poisson model is the foundation of the unsourced random access framework developed by Polyanskiy (2017), which is a major research direction in the CommIT group.
When to Use Poisson vs. Exact Binomial
In modern computing, evaluating the binomial PMF is trivial for moderate (say ). The Poisson approximation is valuable when: (1) is unknown or variable but can be estimated; (2) the model is inherently Poisson (e.g., arrivals from a Poisson process); (3) you need closed-form expressions for further analysis (e.g., deriving capacity formulas that involve sums over Poisson-distributed counts).
Historical Note: Siméon Denis Poisson and the Law of Small Numbers
19th–20th centurySiméon Denis Poisson published his treatise Recherches sur la probabilité des jugements in 1837, where the distribution that bears his name appears as a limit of the binomial. The catchy name "law of small numbers" (Gesetz der kleinen Zahlen) was coined by Ladislaus Bortkiewicz in 1898, who demonstrated the empirical fit on data ranging from Prussian cavalry deaths to children's suicides. The Poisson distribution later became the cornerstone of queueing theory through the work of A. K. Erlang on telephone traffic (1909).
Historical Note: Lucien Le Cam and Approximation Theory
20th centuryLucien Le Cam proved his celebrated inequality in 1960 while working at UC Berkeley. The result was a byproduct of his broader program on the approximation of statistical experiments. Le Cam's inequality remains the standard tool for bounding the error of the Poisson approximation and has been refined by Stein, Chen, and others into the powerful "Stein-Chen method" for Poisson approximation of dependent rare events.
Poisson Limit:
Theorem: Sum of Independent Poisson Random Variables
If and are independent, then .
Poisson counts are closed under addition of independent components. If packets arrive from two independent sources at rates and , the total arrival count is Poisson with rate . This superposition property is fundamental to queueing theory.
Compute the PMF of $X + Y$
$
Simplify using the binomial theorem
\text{Poisson}(\lambda_1 + \lambda_2)\qquad \square$
Common Mistake: Mean Equals Variance Does Not Imply Poisson
Mistake:
Concluding that data is Poisson just because the sample mean and sample variance are approximately equal.
Correction:
The condition is necessary but not sufficient for a Poisson distribution. Many other distributions (e.g., certain negative binomial or compound Poisson distributions) can also have equal mean and variance. A proper goodness-of-fit test (e.g., chi-squared) is needed.
Common Mistake: Applying Poisson Approximation When Is Not Small
Mistake:
Using the Poisson approximation when is, say, 0.3 and .
Correction:
Le Cam's bound gives , which is useless (total variation is at most 1). The Poisson approximation requires . For moderate and large , use the normal approximation (CLT) instead.
Quick Check
If , which distribution best approximates ?
, and is small, so the Poisson approximation is excellent.
Quick Check
Le Cam's inequality bounds the total variation distance between the distribution of a sum of independent Bernoullis and a Poisson by:
Correct. The total variation distance is at most .
Unsourced Random Access and the Poisson User Model
The unsourced random access paradigm introduced by Polyanskiy (2017) models the number of active users in a massive IoT system as a Poisson random variable. This elegant abstraction removes the need to identify individual users and focuses on the fundamental limits of communicating a list of messages. The CommIT group at TU Berlin has been a leading contributor to practical coding schemes for this setting, including slotted transmission and tree-based decoding algorithms.
Poisson distribution
A discrete distribution on with parameter and PMF . Both the mean and variance equal .
Related: binomial distribution
total variation distance
For discrete distributions and on the same space: . It equals the maximum difference over all events .
Related: Poisson distribution
Key Takeaway
The Poisson limit theorem shows that as : the Poisson distribution is the natural model for the count of many independent rare events. Le Cam's inequality quantifies the approximation error. The Poisson distribution's additivity and single-parameter simplicity make it indispensable for modeling packet arrivals, interference events, and user activity in telecommunications.