Finite Probability Spaces and Equally Likely Outcomes
The Classical Model: Symmetry and Equal Likelihood
The classical probability model assigns equal probability to every outcome. It is both the historically oldest probability model (Laplace, 1812) and the one that arises most naturally in information theory and coding: a uniformly chosen codeword, a randomly selected scheduling slot, a uniformly distributed hash input.
The power of the classical model is that it reduces probability to combinatorics: . Its limitation is the assumption of symmetry, which must be justified by the problem's physics or structure. When outcomes are not equally likely — a biased channel, a non-uniform source — one must use the general Kolmogorov framework from Section 1.3.
This section develops the classical model, explores when it applies, and uses it to derive several results that will reappear throughout the book.
Definition: Uniform Probability on a Finite Set
Uniform Probability on a Finite Set
Let be a finite sample space with outcomes. The uniform probability measure on assigns: For any event :
The triple with this is called the classical probability space or equally likely outcomes model.
We verify Kolmogorov's axioms: non-negativity () and normalization () are immediate. Countable additivity reduces to finite additivity for finite : for disjoint , , so .
Example: Uniform Distribution on a Die
Roll a fair die (, uniform). (a) Find . (b) Find . (c) Find . (d) Find .
(a) $\mathbb{P}(\text{even})$
, . .
(b) $\mathbb{P}(\text{outcome} \leq 3)$
, . .
(c) Intersection
(even AND ). .
(d) Union via inclusion-exclusion
. Check: , , .
Example: Random Codebook Selection
Shannon's random coding argument: a binary codebook consists of codewords, each chosen independently and uniformly from . Compute the probability that two specified codewords and () are identical.
Set up the sample space
Each codeword is chosen uniformly from , giving for the joint codebook selection. For the pair : .
Count favorable outcomes
requires both to be the same string: choices. So .
Compute the probability
R < 1/2n \to \inftyR < 1/2$ almost surely has distinct codewords. This is a baby version of Shannon's random coding argument.
Classical Model vs. General Probability Space
| Property | Classical (Uniform) Model | General Kolmogorov Model |
|---|---|---|
| Sample space | Finite | Finite, countable, or uncountable |
| Sigma-algebra | Power set | Any sigma-algebra |
| Probability of outcome | (by symmetry) | Arbitrary with |
| Probability of event | (counting) | Sum or integral over |
| Computing | Combinatorics | Integration or summation |
| When applicable | Symmetric/equally likely outcomes | Always (with specified model) |
| Example | Fair die, random codeword | Rayleigh fading, Gaussian noise |
Why This Matters: Uniform Randomness in Information Theory and Coding
The classical probability model is not just a pedagogical exercise — it is a key structural assumption in information theory. Shannon's channel capacity theorem is proved by showing that a uniformly random codebook achieves a vanishing error probability as the block length , for any rate . The argument rests on two ingredients: the uniform distribution over codewords and the law of large numbers (which we prove in Chapter 11).
Similarly, in MIMO systems, the capacity-achieving input distribution for an i.i.d. Rayleigh fading channel is Gaussian — but the discrete approximation is a uniform distribution over a QAM constellation. Understanding when "all outcomes equally likely" is a valid model is thus essential for interpreting every capacity formula and every simulation result.
Theorem: Hypergeometric Probability
An urn contains balls: are red and are blue. Draw balls without replacement. The probability that exactly of the drawn balls are red is:
The denominator counts all equally likely size- subsets of the balls. The numerator counts subsets with exactly red balls: choose of the red balls ( ways) and of the blue balls ( ways).
Count the sample space
Drawing balls from without replacement gives equally likely outcomes (the drawn subset is unordered).
Count favorable outcomes
To get exactly red: choose from the red balls ( ways), and from the blue balls ( ways). By the multiplication principle: total.
Apply uniform probability
.
Random Coding and the Probabilistic Method in Information Theory
The random coding argument — assigning a uniform probability measure to the codebook ensemble and showing that the average error probability vanishes — is the central proof technique of information theory. It was introduced by Shannon in 1948 and refined in the following decades into a systematic methodology. The probabilistic method here is not just a proof tool but a constructive strategy: a randomly chosen code is likely to be good, even if we cannot exhibit the best code explicitly. This connects directly to the counting arguments in this section: random codebooks are uniform distributions over finite sets, and every bound in the proof uses the inclusion-exclusion and union bound from Section 1.3.
Common Mistake: The Classical Model Requires Genuine Symmetry
Mistake:
A student argues: "There are two outcomes for a biased coin — Heads or Tails — so each has probability 1/2." Or, more subtly: "The channel either makes an error or it doesn't, so the error probability is 1/2."
Correction:
The classical model is valid only when the outcomes are genuinely symmetric — when the physical mechanism treats all outcomes equally. A biased coin has ; the number of outcomes () does not determine their probabilities. For a channel, the error probability is a property of the channel and the signaling scheme, not something we can assign by counting. The classical model applies when: (a) a random experiment is explicitly designed to be fair (shuffled cards, fair dice, uniform random codebook), or (b) physical symmetry enforces equal likelihood (crystal symmetry, isotropic noise).
Uniform Random Scheduling in Wireless Systems
In multi-user wireless systems, a simple and fair scheduling policy is round-robin or uniform random selection: at each time slot, the base station selects one of users uniformly at random to serve. Under this classical model, each user is served with probability per slot.
The probability that a given user is NOT served for consecutive slots is (using ). By the union bound over users, the probability that any user waits more than slots is at most .
This elementary calculation — uniform distribution, union bound, exponential bound — illustrates the practical reach of the tools in this chapter.
- •
Uniform random scheduling is optimal in throughput only under symmetric channels
- •
Proportional fair scheduling (Chapter 11 of Book FSI) adapts weights based on channel quality
Quick Check
A packet of 20 chips contains 5 defective and 15 good ones. An inspector randomly selects 4 chips without replacement. What is the probability that exactly 2 are defective?
By the hypergeometric formula with , , , : .
Key Takeaway
The classical model converts probability to counting, but requires genuine symmetry. It underpins Shannon's random coding argument, uniform scheduling analysis, and birthday-type collision bounds. Whenever a problem says "uniformly at random" or "equally likely," the classical model applies and the calculation reduces to combinatorics.