The Probabilistic Method

Existence Proofs Without Explicit Construction

The probabilistic method is one of the most powerful techniques in combinatorics. The idea is disarmingly simple: to prove that a combinatorial object with some property exists, we define a random experiment over all objects and show that the expected value of some quantity makes it impossible for every object to violate the property.

The argument never constructs the object — it merely proves one must exist. This is existence proof via probability. Paul Erdős pioneered this method in the 1940s and 1950s, and Shannon used it in 1948 to prove the channel coding theorem — perhaps its most consequential application.

Theorem: The Probabilistic Method

Let X\mathcal{X} be a finite set and XX a random variable uniformly distributed over X\mathcal{X}. If E[f(X)]<t\mathbb{E}[f(X)] < t for some function f:XRf: \mathcal{X} \to \mathbb{R}, then there exists at least one xXx \in \mathcal{X} such that f(x)<tf(x) < t.

Equivalently: if P(f(X)t)<1\mathbb{P}(f(X) \geq t) < 1, then P(f(X)<t)>0\mathbb{P}(f(X) < t) > 0, and in particular some xx satisfies f(x)<tf(x) < t.

The average value of ff is below tt; therefore ff cannot exceed tt everywhere — at least one point must be below the average. This trivial observation, applied to the right random experiment, yields non-trivial existence results.

Key Takeaway

The probabilistic method converts a lower bound on E[f(X)]\mathbb{E}[f(X)] into an existence guarantee. The random experiment is a proof device, not a description of how to find the good object.

Historical Note: Erdős and the Birth of the Probabilistic Method

1947–present

Paul Erdős (1913–1996) introduced the probabilistic method in a 1947 paper showing that Ramsey numbers R(k,k)>2k2k/2R(k,k) > \sqrt{2}\cdot k \cdot 2^{k/2} — meaning that large graphs can simultaneously avoid both a clique of size kk and an independent set of size kk, which seems impossible to construct but is easy to prove via a random coloring argument. Erdős had no explicit construction; he only showed that a random coloring succeeds with positive probability. Over the following decades, he and collaborators applied this technique to dozens of seemingly intractable combinatorial problems. The method has since become foundational in coding theory, cryptography, and algorithm design.

Definition:

Random Codebook

A binary random codebook C\mathcal{C} of rate RR and blocklength nn is generated by independently choosing M=2nRM = 2^{nR} codewords c1,,cM{0,1}n\mathbf{c}_1, \ldots, \mathbf{c}_M \in \{0,1\}^n, each component drawn i.i.d.\ from Bernoulli(1/2)\text{Bernoulli}(1/2).

The codebook C\mathcal{C} is a random object. A fixed codebook is any specific realization of this process. Shannon's argument shows that the expected error probability over the random ensemble is small whenever R<CR < C (the channel capacity), proving a good codebook exists.

This definition will be revisited and made precise in Book ITA. Here we emphasize the probabilistic existence argument only.

Theorem: Shannon's Random Coding Existence Argument (Preview)

Consider a binary symmetric channel (BSC) with crossover probability p<1/2p < 1/2. The capacity is C=1Hb(p)C = 1 - H_b(p) bits per channel use, where Hb(p)=plog2p(1p)log2(1p)H_b(p) = -p\log_2 p - (1-p)\log_2(1-p) is the binary entropy. For any rate R<CR < C and ϵ>0\epsilon > 0, there exists a code of rate RR and blocklength nn (sufficiently large) such that the maximum probability of error is at most ϵ\epsilon.

Proof sketch (probabilistic method): Choose a random codebook of rate RR. The expected probability of error over the random ensemble satisfies EC[Pe(C)]    2n(CR)/2  n  0.\mathbb{E}_{\mathcal{C}}[P_e(\mathcal{C})] \;\leq\; 2^{-n(C-R)/2} \;\xrightarrow{n \to \infty}\; 0. Since the expected error is small, there exists a codebook achieving this error probability. \blacksquare

A random code is almost as good as the best possible code when nn is large. The probabilistic method guarantees the existence of good codes without telling us which specific codebook is good. Finding efficient constructive codes that match Shannon's bound (polar codes, LDPC codes) took another 60 years.

,

Connection to Book ITA

The argument above is a sketch of Shannon's 1948 random coding proof. Book ITA (Information Theory and Applications) develops this rigorously: it defines typical sequences and their counting properties (method of types), states and proves the channel coding theorem in full generality for memoryless channels, and analyzes the error exponent (the rate at which Pe0P_e \to 0 as nn \to \infty). The key insight here is the probabilistic method: the existence of a good code follows from the existence of a good average code, without constructing it.

Shannon's Random Coding: Expected Error vs. Rate and Blocklength

The expected probability of error E[Pe]\mathbb{E}[P_e] for a random BSC codebook as a function of rate RR and blocklength nn. Drag the rate slider: for R<CR < C the error decays to zero; for R>CR > C it approaches 1. The phase transition at R=CR = C is the capacity threshold.

Parameters
0.1
500

Example: Tournament Scheduling via the Probabilistic Method

A tournament on nn players is a complete directed graph: every pair of players has exactly one directed edge (the winner of their match). A Hamiltonian path in a tournament is an ordering of all players such that each player beats the next. Prove that every tournament contains a Hamiltonian path (a dominant ranking exists).

Common Mistake: The Probabilistic Method Is Not Constructive

Mistake:

After proving that a good codebook exists via the probabilistic method, one might think we now have a practical coding scheme. In fact, we only know the codebook exists — we have no efficient algorithm to find it.

Correction:

The probabilistic method provides an existence guarantee only. Practical coding schemes (turbo codes, LDPC codes, polar codes) are constructed using algebraic or iterative methods and proven separately to achieve capacity. Book ITA covers polar codes (2009 Erdal Arıkan), which are the first capacity-achieving codes with an efficient construction and polynomial-time encoder/decoder.

Quick Check

The probabilistic method proves existence of a good object by showing:

The expected value of some quality measure is below a threshold

The probability of a good object existing equals 1

An explicit construction algorithm succeeds with high probability

The worst-case object has quality below the threshold

🔧Engineering Note

Why We Do Not Use Random Codes in Practice

Shannon's random coding argument proves that good codes exist, but a uniformly random codebook of rate RR and blocklength nn has 2nR2^{nR} codewords each of length nn bits — the codebook storage alone requires 2nRn2^{nR} \cdot n bits. At rate R=1/2R = 1/2 and n=128n = 128 (a modest blocklength for 5G NR), storing the codebook would require 2641282^{64} \cdot 128 bits 1021\approx 10^{21} bits — far beyond any imaginable storage. Practical codes (polar, LDPC, convolutional) have algebraic structure so the encoder and decoder require only O(nlogn)O(n \log n) operations and O(n)O(n) storage.

Practical Constraints
  • Polar codes (3GPP NR control channels) achieve capacity with O(nlogn)O(n \log n) encoder and decoder complexity

  • LDPC codes (3GPP NR data channels) are capacity-approaching with sparse parity-check matrices, O(n)O(n) decoder

  • 5G NR uses polar codes for PBCH, PDCCH; LDPC for PDSCH/PUSCH (TS 38.212)

📋 Ref: 3GPP TS 38.212