Chapter Summary

Chapter Summary

Key Points

  • 1.

    The multivariate Gaussian N(μ,Σ)\mathcal{N}(\boldsymbol{\mu}, \boldsymbol{\Sigma}) is fully specified by its mean and covariance. The PDF involves the precision matrix Σ1\boldsymbol{\Sigma}^{-1} in the exponent, and the constant-density contours are ellipsoids whose axes align with the eigenvectors of Σ\boldsymbol{\Sigma}.

  • 2.

    Marginals are Gaussian. If XN(μ,Σ)\mathbf{X} \sim \mathcal{N}(\boldsymbol{\mu}, \boldsymbol{\Sigma}) is partitioned into (X1,X2)(\mathbf{X}_1, \mathbf{X}_2), then X1N(μ1,Σ11)\mathbf{X}_1 \sim \mathcal{N}(\boldsymbol{\mu}_1, \boldsymbol{\Sigma}_{11}) — simply read off the corresponding block.

  • 3.

    Conditionals are Gaussian with Schur complement formulas. X1X2=x2N(μ12,Σ12)\mathbf{X}_1 | \mathbf{X}_2 = \mathbf{x}_2 \sim \mathcal{N}(\boldsymbol{\mu}_{1|2}, \boldsymbol{\Sigma}_{1|2}) where μ12=μ1+Σ12Σ221(x2μ2)\boldsymbol{\mu}_{1|2} = \boldsymbol{\mu}_1 + \boldsymbol{\Sigma}_{12}\boldsymbol{\Sigma}_{22}^{-1}(\mathbf{x}_2 - \boldsymbol{\mu}_2) is affine in x2\mathbf{x}_2 and Σ12=Σ11Σ12Σ221Σ21\boldsymbol{\Sigma}_{1|2} = \boldsymbol{\Sigma}_{11} - \boldsymbol{\Sigma}_{12}\boldsymbol{\Sigma}_{22}^{-1}\boldsymbol{\Sigma}_{21} does not depend on x2\mathbf{x}_2.

  • 4.

    Affine transformations preserve Gaussianity. AX+bN(Aμ+b,AΣAT)\mathbf{A}\mathbf{X} + \mathbf{b} \sim \mathcal{N}(\mathbf{A}\boldsymbol{\mu} + \mathbf{b}, \mathbf{A}\boldsymbol{\Sigma}\mathbf{A}^T). Every linear combination of Gaussian components is scalar Gaussian.

  • 5.

    Uncorrelated \Longleftrightarrow independent for Gaussians. This is the defining structural advantage of the Gaussian family: second-order analysis (decorrelation, PCA, whitening) achieves full statistical independence.

  • 6.

    The whitening transform maps X\mathbf{X} to WN(0,I)\mathbf{W} \sim \mathcal{N}(\mathbf{0}, \mathbf{I}). It is the multivariate analogue of standardization and is the first step in many detection and estimation algorithms.

  • 7.

    Chi-squared and Wishart distributions arise from quadratic functions of Gaussians. The Mahalanobis distance (Xμ)TΣ1(Xμ)χn2(\mathbf{X} - \boldsymbol{\mu})^T\boldsymbol{\Sigma}^{-1}(\mathbf{X} - \boldsymbol{\mu}) \sim \chi^2_n. The sample covariance matrix follows a Wishart distribution.

  • 8.

    The proper complex Gaussian CN(μ,Σ)\mathcal{CN}(\boldsymbol{\mu}, \boldsymbol{\Sigma}) models baseband noise and Rayleigh fading. Circular symmetry (vanishing pseudo-covariance) means the distribution is invariant to phase rotation. All real Gaussian properties carry over.

Looking Ahead

Chapter 9 develops generating functions and transforms — the moment generating function and characteristic function — as systematic tools for analyzing sums and limits of random variables. The multivariate characteristic function derived here will be the starting point for proving the multivariate Central Limit Theorem in Chapter 11, which explains why the Gaussian distribution appears so ubiquitously.