Affine Transformations and the Whitening Transform
Why Affine Transformations Matter
In engineering, we constantly transform random vectors: a receive filter multiplies the observation by a matrix, a decoder adds a bias, a pre-whitening step decorrelates the input. The fundamental question is: if is Gaussian, what can we say about ? The answer is that is also Gaussian, and its mean and covariance follow directly from the affine map. This closure property is the reason Gaussian models are so tractable.
Theorem: Affine Transformations Preserve Gaussianity
Let and let , where and . Then
A linear map rotates and stretches the Gaussian "cloud," while the translation shifts the center. The covariance transforms as a quadratic form because it involves the outer product of .
Compute the mean
.
Compute the covariance
.
Gaussianity via characteristic function
The CF of is , which is the CF of .
Every Linear Combination Is Scalar Gaussian
An immediate corollary: for any fixed , the scalar is distributed as .
In fact, this property characterizes the Gaussian: a random vector is Gaussian if and only if every linear combination of its components is a scalar Gaussian (or degenerate). This is the Cramér–Wold device applied to the Gaussian family.
Definition: Whitening Transform
Whitening Transform
Let with . The whitening transform is
where is any matrix satisfying . By the affine transformation theorem,
Common choices for :
- (eigendecomposition),
- (inverse Cholesky factor, yielding a causal whitening filter).
The whitening transform is the multivariate analogue of standardization . The resulting has independent standard Gaussian components.
Whitening transform
A linear transformation that maps a random vector with covariance to one with covariance (uncorrelated, unit-variance components). For Gaussians, whitening produces independent components.
Related: Covariance matrix
Whitening Transform Demo
Start with a correlated bivariate Gaussian and apply the whitening transform. The left panel shows the original elliptical contours; the right panel shows the whitened circular contours. Adjust to see how stronger correlation leads to more dramatic reshaping.
Parameters
Example: Whitening a Bivariate Gaussian
Let with . Find the whitening transform using the eigendecomposition.
Eigendecompose $\ntn{covmat}$
The eigenvalues are , , with eigenvectors and .
Compute $\ntn{covmat}^{-1/2}$
$
Verify
has covariance . The whitened components are i.i.d. .
Definition: Karhunen-Loeve Expansion
Karhunen-Loeve Expansion
Let be the eigendecomposition with and . For , define . Then
where are independent. This is the Karhunen-Loeve (KL) expansion: is a sum of deterministic eigenvector "modes" with independent random amplitudes .
The KL expansion is the probabilistic analogue of PCA (principal component analysis). The first terms capture the most variance, giving the best rank- approximation to in the MSE sense.
Principal Axes of a 2D Gaussian
Visualize the eigendecomposition of : the principal axes (eigenvectors) and their lengths (). Rotate and scale the covariance by adjusting parameters.
Parameters
Common Mistake: The Whitening Transform Is Not Unique
Mistake:
Assuming there is a single whitening matrix for a given .
Correction:
Any matrix satisfying is a valid whitening transform. Two common choices are the eigendecomposition-based and the Cholesky-based . The Cholesky version is preferred in filtering because it is "causal" — depends only on .
Whitening as a Pre-Processing Step
In detection and estimation, a common first step is to whiten the observation: given with , multiply both sides by to get where . This reduces colored-noise problems to the white-noise case, where standard matched filtering and LMMSE formulas apply directly.
Quick Check
If and , what is ?
.