Repeated Independent Trials
Building Distributions from Independent Trials
The most widely used probability distributions in communications β binomial, geometric, negative binomial β arise naturally from repeating a simple binary experiment (Bernoulli trial) multiple times, independently. Rather than postulating these distributions axiomatically, we derive them from first principles: a single parameter (the success probability) generates a whole family of distributions by asking different questions about the repeated experiment.
This derivation is important because it shows exactly which independence assumptions underlie each distribution. When those assumptions fail β bursty errors, correlated channel states β these distributions no longer apply and must be replaced.
Definition: Bernoulli Trial and Bernoulli Sequence
Bernoulli Trial and Bernoulli Sequence
A Bernoulli trial is an experiment with exactly two outcomes: success (S) and failure (F), with and for some .
A Bernoulli sequence is an infinite sequence of independent, identically distributed Bernoulli trials where each (success) with probability and (failure) with probability .
The i.i.d. assumption is the key structural property. In a Bernoulli sequence, the outcome of any trial provides no information about any other trial β neither past nor future. This is the memoryless property at the sequence level.
Theorem: Binomial Distribution
In a Bernoulli sequence with parameter , let be the number of successes in trials. Then has the binomial distribution with probability mass function The mean and variance are and .
Any specific sequence of successes and failures has probability . The factor counts the number of such sequences. Since they are mutually exclusive and exhaustive, the probabilities sum to by the binomial theorem.
Probability of a specific sequence
By independence, a specific ordered outcome with S's and F's has probability (multiply the probabilities of each trial).
Count the sequences
The number of ordered sequences of length containing exactly successes equals the number of ways to choose positions from for the successes: .
Sum over mutually exclusive sequences
The event is the union of all specific sequences, which are mutually exclusive. Therefore:
Mean and variance
By linearity of expectation, . By independence, (since for a Bernoulli variable).
Definition: Geometric Distribution
Geometric Distribution
In a Bernoulli sequence with parameter , let be the waiting time until the first success: . Then has the geometric distribution with PMF and mean .
The geometric distribution is the unique discrete distribution with the memoryless property: . Knowing that no success has occurred in the first trials does not change the distribution of the remaining waiting time. This mirrors the continuous memoryless property of the exponential distribution.
Theorem: Memoryless Property of the Geometric Distribution
Let . For any integers : Conversely, the geometric distribution is the only discrete distribution on with this property.
Compute the tail probability
. (Geometric series: .)
Verify the memoryless property
$
Uniqueness (sketch)
Suppose for all . Setting , we have with and . The only such function on the non-negative integers is for some . Setting recovers the geometric distribution.
Definition: Negative Binomial Distribution
Negative Binomial Distribution
In a Bernoulli sequence with parameter , let be the waiting time until the -th success. Then has the negative binomial distribution with PMF Mean: . Variance: .
The name "negative binomial" comes from the fact that the PMF arises from the binomial series with negative exponent. The special case is the geometric distribution. The sum of independent Geom random variables has the NegBin distribution.
Example: Binomial Model for Bit Error Count
A BPSK link has bit error probability . A packet contains bits. Compute (a) the expected number of bit errors, (b) the probability of zero errors, and (c) the probability of more than 2 errors.
Model
Assume errors occur independently (Bernoulli sequence). The number of errors .
(a) Expected errors
.
(b) Zero errors
(1-1/n)^n \approx e^{-1}n$.)
(c) More than 2 errors
. Since is moderate and is large, the Poisson approximation gives: . So .
Binomial Distribution from Bernoulli Trials
Visualize the PMF and how it evolves as and vary. Notice the approach to a Gaussian bell curve as increases (central limit theorem preview).
Parameters
Probability Tree for Binomial Trials
Monty Hall Simulation
Simulate the Monty Hall problem: rounds, switch vs. stay strategy. Watch the empirical win rate converge to the theoretical probabilities (switch) and (stay). This is an application of Bayes' theorem: after the host reveals a goat, the posterior probability of the unchosen door having the car increases.
Parameters
Packet Error Rate and the Binomial Model
The binomial model for packet errors assumes independent bit errors β a valid approximation when the channel uses interleaving to break burst correlation. The packet error rate (PER) for a packet of bits and bit error rate is: Without forward error correction (FEC), a -bit packet at has . With a rate- convolutional code ( coded bits) and interleaving, the effective after decoding can fall below , reducing PER to .
- β’
LTE/5G NR target PER is at SINR thresholds specified in TS 38.214
- β’
Independent error assumption requires interleaver depth coherence time
- β’
HARQ retransmissions can recover from occasional packet errors at the system level
Binomial Distribution
The number of successes in independent Bernoulli trials. PMF: . Mean , variance .
Related: Bernoulli Trial and Bernoulli Sequence, Geometric Distribution, Poisson Approximation
Geometric Distribution
Waiting time until the first success in a Bernoulli sequence. PMF: . Mean . The unique memoryless distribution on .
Related: Bernoulli Trial and Bernoulli Sequence, Memoryless Property, Exponential Distribution
Quick Check
A transmission system has word error rate . In independent transmissions, what is the probability of exactly 2 errors?
.
Key Takeaway
Three distributions, one experiment. Bernoulli trials generate the binomial (how many successes in trials?), geometric (how long until the first success?), and negative binomial (how long until the -th success?) distributions β all from the same parameter . The geometric distribution is the only discrete memoryless distribution, paralleling the exponential distribution in continuous time. In error probability analysis, these distributions quantify performance under the i.i.d. error model.