The Karhunen-Loève Expansion
The Optimal Basis for a Random Process
Given a random process on a finite interval , we want to represent it as a series for some basis functions and random coefficients . But which basis is best? The Fourier basis is convenient but not adapted to the process statistics. The point is that the Karhunen-Loève (KL) expansion chooses the basis that diagonalizes the autocorrelation operator — the eigenfunctions of . This makes the coefficients uncorrelated (and independent for Gaussian processes), concentrates the maximum energy in the fewest terms, and provides the optimal finite-dimensional approximation in the mean-square sense. The KL expansion is the continuous-time analogue of principal component analysis (PCA).
Definition: The Karhunen-Loève Expansion
The Karhunen-Loève Expansion
Let be a zero-mean, finite-variance random process on with autocorrelation . The Karhunen-Loève (KL) expansion of is where:
- The functions are the orthonormal eigenfunctions of the autocorrelation kernel, satisfying the Fredholm integral equation
- The eigenvalues are real and non-negative.
- The random coefficients are .
- The are uncorrelated: .
- For Gaussian , the are independent: .
The convergence is in the mean-square sense: for each . For Gaussian processes, convergence also holds uniformly on with probability one under mild regularity.
Karhunen-Loève Expansion
A series representation where are eigenfunctions of the autocorrelation kernel and are uncorrelated random coefficients with variance equal to the corresponding eigenvalues.
Related: Fredholm Integral Equation, Principal Component Analysis (PCA)
Fredholm Integral Equation
The eigenvalue problem for an integral operator with kernel . In the KL expansion, the kernel is the autocorrelation function.
Related: Karhunen-Loève Expansion
Principal Component Analysis (PCA)
The discrete analogue of the KL expansion. For a random vector with covariance , PCA represents in the eigenbasis of , yielding uncorrelated components ordered by variance.
Related: Karhunen-Loève Expansion
Theorem: Optimality of the KL Expansion (Minimum Mean-Square Truncation Error)
Among all orthonormal expansions with uncorrelated coefficients, the KL expansion minimizes the mean-square truncation error:
Specifically, , and no other -term orthonormal expansion achieves a smaller error.
The KL basis concentrates the process energy into the first few coefficients by construction: the eigenvalues are ordered , so the first terms capture the maximum possible energy. Any other basis would "spread" energy more evenly across coefficients, requiring more terms for the same approximation quality.
Express the total energy
The total average energy over is where the last equality follows from Mercer's theorem (the eigenvalue expansion of the kernel's trace).
Compute the truncation error
The -term approximation has energy . The residual has energy . Therefore .
Prove optimality
Let be any other orthonormal basis with coefficients . The energy in the first components is . By a variational argument (or directly by the Courant-Fischer min-max theorem applied to the integral operator), is achieved when (the KL eigenfunctions), and the maximum equals . Hence the truncation error is minimized by the KL basis.
Example: KL Expansion of the Wiener Process on
Find the KL expansion of the Wiener process on , which has autocorrelation .
Set up the Fredholm equation
We need . Split the integral at :
Convert to a differential equation
Differentiating twice with respect to (using Leibniz's rule), we get with boundary conditions (from the integral equation at ) and (from differentiating once and evaluating at ).
Solve the ODE
The general solution of is with . The boundary condition gives . The condition gives , so for .
Write the eigenvalues and eigenfunctions
W(t) = \sum_{n=1}^{\infty} Z_n \phi_n(t)Z_n \sim \mathcal{N}(0, \lambda_n)\sum_n \lambda_n = \sigma^2 T\mathbb{E}[W(T)^2] = \sigma^2 T$.
Karhunen-Loève Expansion: Convergence vs. Number of Terms
Visualize the KL expansion of a Wiener process on . As more eigenfunctions are included, the expansion approximates the true realization more closely. Observe how the eigenvalues decay (lower plot) and how the truncation error decreases.
Parameters
Karhunen-Loève Basis Functions Building Up a Random Realization
The KL Expansion in Detection Theory
The KL expansion transforms the continuous-time detection problem "test vs. for " into an equivalent discrete problem in the KL coefficients. Under , the KL coefficients are ; under , they are where . Since the are independent for Gaussian noise, the likelihood ratio factors, and we recover the matched filter as the sufficient statistic. This is the rigorous justification of the matched filter derived heuristically in Ch. 15.
KL Expansion vs. Fourier Series
The Fourier series and the KL expansion are both orthonormal expansions of a process on , but they differ in a crucial way:
- The Fourier basis is fixed and independent of the process. The coefficients are generally correlated unless is WSS with specific structure.
- The KL basis is adapted to the process statistics. The coefficients are always uncorrelated (by construction).
For a WSS process on a long interval, the KL eigenfunctions approach the Fourier exponentials, and the eigenvalues approach the PSD samples . This is the connection between the KL expansion and the Wiener-Khinchin theorem.
Definition: Mercer's Theorem
Mercer's Theorem
If is a continuous, positive semi-definite kernel on , then it admits the eigenvalue expansion where convergence is absolute and uniform. In particular, so the total energy equals the sum of eigenvalues (the trace of the operator).
Definition: KL Expansion of WSS Processes on Large Intervals
KL Expansion of WSS Processes on Large Intervals
For a WSS process on with large, the KL eigenfunctions are approximately complex exponentials with , and the eigenvalues are approximately PSD samples: where .
This means for large , the KL expansion coincides with the Fourier expansion, and the eigenvalue distribution converges to the PSD.
Karhunen-Loève vs. Fourier Expansion
| Property | KL Expansion | Fourier Series |
|---|---|---|
| Basis functions | Eigenfunctions of — adapted to the process | Complex exponentials — fixed |
| Coefficients | Always uncorrelated (independent for Gaussian) | Generally correlated |
| Optimality | Minimizes -term m.s. truncation error | Not optimal in general |
| Computation | Requires solving a Fredholm integral equation | FFT — fast and simple |
| Large limit | Approaches Fourier for WSS processes | Approaches Fourier (tautologically) |
| Non-stationary processes | Handles naturally | Not adapted — poor convergence |
Historical Note: Karhunen and Loève
1947-1960sKari Karhunen (1947, Finland) and Michel Loève (1948, France) independently discovered the expansion that bears their names. Karhunen was a student of Rolf Nevanlinna at the University of Helsinki, and his original paper was in Finnish — one reason the result was initially less known in the West. Loève, working in France and later at UC Berkeley, developed the expansion within his comprehensive theory of second-order processes. The KL expansion became central to communication theory through the work of David Slepian at Bell Labs, who in the 1960s computed the KL eigenfunctions for bandlimited processes (the prolate spheroidal wave functions), establishing the mathematical theory of time-frequency concentration.
Karhunen-Loève Channel Representation for Massive MIMO
The CommIT group used KL-type decompositions of the spatial channel covariance matrix to develop the Joint Spatial Division and Multiplexing (JSDM) framework for massive MIMO. The idea is to group users by their channel covariance eigenspaces — effectively by their KL bases — and serve each group with a pre-beamformer that projects onto the dominant eigenmodes. This two-stage beamforming approach (statistical pre-beamformer + instantaneous beamformer) achieves near-optimal massive MIMO capacity with only reduced-dimension CSI feedback. The covariance eigenmodes are precisely the spatial KL basis functions, and the eigenvalues determine how many spatial degrees of freedom each user group occupies. The JSDM framework demonstrates that the KL expansion is not merely a theoretical tool but a practical architecture for next-generation wireless systems.
Practical KL Truncation for Signal Compression
In practice, one truncates the KL expansion to terms, discarding eigenmodes with below a threshold. The fraction of energy captured is . For many processes of interest (exponential autocorrelation, bandlimited processes), the eigenvalues decay rapidly, so with the nominal dimension . This is the principle behind transform coding, PCA-based compression, and reduced-rank signal processing. In massive MIMO, the rapid decay of spatial covariance eigenvalues means that only beams are needed to capture most of the channel energy.
- •
Eigenvalue computation for large correlation matrices is — Krylov methods help
- •
Non-stationary processes require recomputing eigenfunctions over time
Quick Check
In the KL expansion of a Gaussian process, the coefficients are:
Independent Gaussian random variables with .
Uncorrelated but not necessarily independent.
i.i.d. Gaussian random variables.
For a Gaussian process, the KL coefficients are uncorrelated (by the general KL theory) and jointly Gaussian (since they are linear functionals of a Gaussian process). Uncorrelated + jointly Gaussian = independent. Each has variance .
Common Mistake: The KL Expansion Is Not Limited to WSS Processes
Mistake:
Assuming the KL expansion applies only to wide-sense stationary processes.
Correction:
The KL expansion applies to any finite-variance process on a bounded interval , whether stationary or not. The autocorrelation kernel need not be a function of alone. For non-stationary processes, the eigenfunctions are not sinusoids but adapted to the specific correlation structure. The WSS case is special only in that the eigenfunctions approximate Fourier exponentials for large .
Key Takeaway
The Karhunen-Loève expansion provides the optimal orthonormal representation of a random process: it diagonalizes the autocorrelation operator, yielding uncorrelated (independent for Gaussian) coefficients ordered by decreasing variance. The -term KL approximation minimizes the mean-square truncation error among all -dimensional projections — this is PCA for random processes. For WSS processes on large intervals, the KL basis converges to the Fourier basis and the eigenvalues converge to the PSD, unifying the spectral and KL viewpoints.