Marginals and Conditionals
Extracting Structure from the Joint
One of the most powerful features of the Gaussian family is that both marginals and conditionals remain Gaussian, with parameters given by simple formulas involving the block structure of . The conditional mean formula is the foundation of LMMSE estimation — and for Gaussians, the LMMSE estimator is the MMSE estimator.
Theorem: Marginals of a Gaussian Are Gaussian
Let with the partition
Then the marginal distribution of is
Marginalizing (integrating out ) simply "reads off" the relevant block of the mean and covariance. No Schur complement or matrix inversion is needed for marginals — only for conditionals.
Integrate out $\mathbf{X}_2$
Write the joint PDF and integrate over . Because the exponent is quadratic in , the integral is a Gaussian integral that evaluates to a constant times the marginal Gaussian PDF.
Alternative via characteristic function
Set in the joint characteristic function . This gives , which is the CF of .
Theorem: Conditional Distribution of a Gaussian Vector
With the same partition as above and , the conditional distribution of given is
where
The matrix is the Schur complement of in .
Two remarkable properties: (1) the conditional mean is an affine function of , and (2) the conditional covariance does not depend on . Property (1) means the MMSE estimator of given is linear — the general MMSE problem reduces to LMMSE in the Gaussian case. Property (2) means the estimation uncertainty is the same regardless of the observed value.
Start from $f_{\mathbf{X}_1|\mathbf{X}_2} = f_{\mathbf{X}_1, \mathbf{X}_2} / f_{\mathbf{X}_2}$
Write the joint PDF using the block partition of . Subtract the marginal PDF of in the exponent.
Use the block matrix inversion identity
The inverse of in block form yields
Complete the square
After substituting the block inverse, the quadratic form in becomes
plus terms depending only on . These cancel with , leaving the Gaussian PDF with mean and covariance .
Schur complement
For a block matrix with invertible, the Schur complement of is . It governs the conditional covariance in the Gaussian conditional distribution.
Related: Covariance matrix
Example: Conditional Distribution for the Bivariate Gaussian
Let with . Find the conditional distribution of given .
Apply the conditional formulas
X_1 \mid X_2 = x_2 ;\sim; \mathcal{N}(\rho x_2,; 1 - \rho^2)$.
Interpretation
The conditional mean is a linear function of with slope — this is the regression line . The conditional variance is smaller than the marginal variance 1 whenever : observing reduces our uncertainty about .
Conditional Distribution Visualizer
Visualize the conditional PDF for a bivariate Gaussian. Drag the slider to change the observed value and watch the conditional density shift while its width (determined by ) stays constant.
Parameters
Common Mistake: The Conditional Variance Does Not Depend on the Observation
Mistake:
Expecting the conditional variance to change depending on the specific observed value .
Correction:
For Gaussian vectors, the conditional covariance depends only on the covariance structure, not on the realized observation. This is a special property of the Gaussian — for most other distributions, the conditional variance does depend on the conditioning value.
Computing the Schur Complement Efficiently
In practice, computing should never involve forming explicitly. Instead, solve the linear system (e.g., via Cholesky factorization of ) and then compute . This is numerically more stable and has complexity rather than the of inverting the full matrix.
Quick Check
For with , what is ?
.