Random Variables and Distributions
Why Random Variables?
A probability space gives us the formal machinery to talk about random experiments, but the sample space can be unwieldy --- outcomes might be waveforms, sequences of channel states, or abstract protocol events. To compute, we need numbers.
A random variable is the bridge: it maps every outcome to a real number, converting abstract randomness into something we can integrate, differentiate, and optimise.
In wireless communications the idea is immediate. The received baseband sample at a single antenna is
where is the transmitted symbol (deterministic, once the codebook is fixed) and is additive noise --- a Gaussian random variable. Understanding the distribution of lets us compute bit-error rates, design detectors, and prove capacity theorems. Every subsequent chapter builds on this section's definitions.
Definition: Random Variable
Random Variable
Let be a probability space. A random variable is a measurable function
meaning that for every Borel set the pre-image belongs to the -algebra .
In particular, the set is an event (element of ) for every , so is well-defined.
The measurability condition is automatically satisfied for every function encountered in engineering practice. Its role is to prevent pathological constructions that would make probabilities undefined.
We write "" (uppercase italic) for a random variable and "" (lowercase italic) for a particular realisation.
Definition: Cumulative Distribution Function (CDF)
Cumulative Distribution Function (CDF)
The cumulative distribution function (CDF) of a random variable is the function defined by
Properties. Every CDF satisfies:
- Right-continuity: is right-continuous at every point, i.e., .
- Monotonicity: is non-decreasing: .
- Boundary limits: and .
Conversely, any function satisfying properties 1--3 is the CDF of some random variable (the Skorokhod construction).
The CDF completely characterises the distribution of . Two random variables with the same CDF are said to be equal in distribution, written .
Definition: Probability Density Function (PDF)
Probability Density Function (PDF)
A random variable is called continuous (or, more precisely, absolutely continuous) if there exists a non-negative function such that
The function is called the probability density function (PDF) of . At every point where is differentiable,
Properties.
- for all .
- .
- .
The PDF is not a probability: can exceed . Only when integrated over an interval does it yield a probability.
Definition: Probability Mass Function (PMF)
Probability Mass Function (PMF)
A random variable is discrete if it takes values in a countable set . Its probability mass function (PMF) is
Properties.
- for all .
- .
The CDF of a discrete random variable is a staircase function with jumps of size at each point .
In digital communications, discrete random variables arise naturally: the transmitted bit , the number of errors in a codeword, and the number of packet arrivals in a time slot are all discrete.
Definition: Expectation (Mean)
Expectation (Mean)
The expectation (or mean) of a random variable is defined as
for continuous (provided the integral converges absolutely), and
for discrete . We also write .
LOTUS (Law of the Unconscious Statistician). For any (measurable) function ,
without needing to derive the distribution of first.
Linearity of expectation. For any random variables and (not necessarily independent) and constants ,
LOTUS is one of the most practically useful results in probability. For example, the average received power is , which LOTUS lets us compute directly from the distribution of without first finding the distribution of .
Definition: Variance and Standard Deviation
Variance and Standard Deviation
The variance of a random variable is
where . We write .
The standard deviation is .
Properties.
- , with equality iff a.s.
- for constants .
- If and are independent, then .
The identity is the computational workhorse. In noise analysis, if is zero-mean (), then , which is the noise power.
Theorem: Properties of Expectation
Let and be random variables on with finite expectations, and let . Then:
-
Linearity: .
-
Monotonicity: If almost surely, then .
-
Triangle inequality: .
-
Product rule for independent RVs: If and are independent, then .
Linearity is the single most important property of expectation and holds without any independence assumption. It is the reason we can decompose the mean of a sum of interference terms without knowing their joint distribution.
Monotonicity formalises the obvious: if one random variable always dominates another, its average must be at least as large.
The product rule for independent RVs is not a consequence of linearity; it requires the factorisation of the joint PDF.
For linearity, use the definition of expectation and linearity of integration.
For monotonicity, note that a.s. implies .
For the product rule, write and factor the joint PDF.
Linearity
For continuous random variables,
Splitting the integral by linearity of integration:
The first double integral equals (integrate out to get the marginal), the second equals , and the third equals . Hence .
Monotonicity
Define . Then a.s., so for . Therefore
and by linearity , i.e., .
Product rule for independent RVs
If and are independent, their joint PDF factors: . Then
Definition: Bernoulli Distribution
Bernoulli Distribution
A random variable has the Bernoulli distribution with parameter , written , if
Mean: .
Variance: .
The Bernoulli distribution is the natural model for a single bit decision. In a binary symmetric channel with crossover probability , the error indicator takes the value when the received bit differs from the transmitted bit.
Definition: Binomial Distribution
Binomial Distribution
A random variable has the binomial distribution with parameters and , written , if
Mean: .
Variance: .
If a block of coded bits is sent through a BSC with crossover probability , and errors are independent across bits, then the total number of bit errors . The binomial distribution thus governs block error analysis for memoryless channels.
Definition: Poisson Distribution
Poisson Distribution
A random variable has the Poisson distribution with rate parameter , written , if
Mean: .
Variance: .
The Poisson distribution arises as the limit of when , , and . In telecommunications it models the number of packet arrivals in a time slot, the number of interferers in a stochastic geometry model, and the number of photons detected in an optical receiver.
Definition: Continuous Uniform Distribution
Continuous Uniform Distribution
A random variable has the uniform distribution on , written , if its PDF is
Mean: .
Variance: .
In wireless communications, the carrier phase offset models the fact that the receiver has no a priori knowledge of the absolute phase. This assumption underlies non-coherent detection and is implicit in the derivation of the Rayleigh fading model.
Definition: Exponential Distribution
Exponential Distribution
A random variable has the exponential distribution with rate parameter , written , if
CDF: for .
Mean: .
Variance: .
Memoryless property. The exponential is the only continuous distribution with the memoryless property:
The exponential distribution models inter-arrival times in a Poisson process. In teletraffic engineering, the time between successive packet arrivals at a router is often modelled as . Additionally, in Rayleigh fading the instantaneous SNR is exponentially distributed when .
Definition: Gaussian (Normal) Distribution
Gaussian (Normal) Distribution
A random variable has the Gaussian (or normal) distribution with mean and variance , written , if its PDF is
Mean: .
Variance: .
The standard normal is obtained by setting , . Any Gaussian can be standardised: .
The Q-function. The tail probability of the standard normal is
where is the complementary error function. This function appears in virtually every error-probability expression in digital communications.
The Gaussian distribution owes its central role to the Central Limit Theorem (Section 2.5): the sum of many independent, identically distributed random variables converges in distribution to a Gaussian, regardless of the individual distributions. Thermal noise, which is the aggregate effect of countless electron vibrations, is therefore accurately modelled as Gaussian.
The complex Gaussian used for baseband noise has independent real and imaginary parts, each and respectively, so that the total power (variance) is .
Gaussian PDF and CDF Explorer
Visualise how the Gaussian PDF and CDF change as you vary the mean and standard deviation . Use the dropdown to overlay Rayleigh, Ricean, or Nakagami- densities --- these fading distributions, derived from the Gaussian in Section 2.3, are previewed here for comparison.
Parameters
Example: Bit-Error Rate of BPSK in AWGN
A binary phase-shift keying (BPSK) system transmits with equal probability over an AWGN channel. The received signal is
The receiver decides if and if . Derive the bit-error rate (BER) as a function of .
Condition on transmitted symbol
By symmetry of the constellation and the decision rule,
Condition on :
Standardise the Gaussian
Since , let . Then
Express in terms of the Q-function
Using the symmetry of the standard normal, :
This is the fundamental BPSK BER formula. At (about in linear scale), .
Equivalent erfc form
Using :
Both forms are widely used; the Q-function form is standard in information-theoretic analyses, while the erfc form is common in link-budget calculations and simulation scripts.
The Gaussian Distribution: Shape, Scale, and the 68-95-99.7 Rule
Why This Matters: The Q-Function and Error Probability in Digital Communications
The Q-function is the universal currency for error-rate computation in AWGN channels. For any modulation scheme whose conditional error event reduces to a Gaussian tail probability, the BER or symbol-error rate (SER) can be expressed in terms of .
Examples:
| Modulation | Error probability |
|---|---|
| BPSK | |
| QPSK | |
| -QAM |
The Q-function decreases faster than any polynomial but slower than a pure exponential. Its asymptotic approximation for large is
which is useful for high-SNR analysis and diversity-order calculations.
Relation to erfc. The Q-function and complementary error function are related by
Most numerical libraries provide directly, making this conversion essential for simulation code.
See full treatment in Chapter 5Ψ Section 2
Quick Check
If for (and for ), what value of makes a valid PDF?
We need . Using the substitution , : . Hence , giving . (This is, in fact, the Rayleigh PDF with .)
Quick Check
Let . What is ?
By linearity, . Note that the variance does not affect the mean calculation.
Quick Check
If , what is ?
. Constants shift the mean but do not affect spread, and scaling by scales the variance by . Hence .
Random variable
A measurable function mapping outcomes of a random experiment to real numbers. Enables computation of probabilities, expectations, and distributions.
Related: Random Variable, Cumulative Distribution Function (CDF), Probability Density Function (PDF)
Probability density function (PDF)
A non-negative function whose integral over any interval gives . It is the derivative of the CDF wherever the derivative exists.
Related: Probability Density Function (PDF), Cumulative Distribution Function (CDF)
Cumulative distribution function (CDF)
The function , which completely characterises the distribution of a random variable. It is non-decreasing, right-continuous, with limits and .
Related: Cumulative Distribution Function (CDF), Probability Density Function (PDF)
Expectation
The weighted average (continuous) or (discrete). A linear functional of the random variable's distribution.
Related: Expectation (Mean), Properties of Expectation
Variance
, a measure of the spread of around its mean. Equals . Units are the square of the units of .
Related: Variance and Standard Deviation
Q-function
The tail probability of the standard normal distribution: where . Ubiquitous in error-probability expressions for digital modulation over AWGN channels.
Related: Gaussian (Normal) Distribution, The Q-Function and Error Probability in Digital Communications, Bit-Error Rate of BPSK in AWGN
Key Takeaway
The core message of this section in three bullets:
-
Random variables turn experiments into numbers. The measurable function is the formal link between abstract probability spaces and the integrals, expectations, and variances that drive engineering design.
-
Six distributions cover most of wireless. Bernoulli and binomial for error counting, Poisson for arrivals, uniform for unknown phase, exponential for inter-arrival times and Rayleigh SNR, and Gaussian for thermal noise --- these six appear on almost every page of a communications textbook.
-
The Gaussian is the workhorse noise model. Because thermal noise is the sum of myriad independent microscopic contributions, the Central Limit Theorem guarantees it is Gaussian. The Q-function, the tail integral of the standard normal, is the single most important special function in digital communications: every BER formula you will encounter in Chapters 5--8 is built from it.
Common Mistake: Forgetting the in the Q-function vs erfc Relationship
Mistake:
Writing without the factor of , or equivalently instead of .
This error is extremely common in homework and even published papers. It leads to BER predictions that are off by exactly a factor of , which can silently propagate through simulation comparisons.
Correction:
The correct relations are:
A useful sanity check: and . Without the factor of one would get , which is clearly wrong since it would imply for a symmetric distribution.
Tip: Always verify your formula at as a quick consistency test.
Common Mistake: Interpreting the PDF Value as a Probability
Mistake:
Claiming that is the probability that . For continuous random variables, for every individual point. Yet it is common to see statements like "the probability of receiving is ."
Correction:
The PDF is a density, not a probability. Only when integrated over an interval does it yield a probability:
Informally, is the probability that falls in an infinitesimally small interval around . In particular, can exceed (e.g., has on its support).