Chapter Summary

Chapter 7 Summary

Key Points

  • 1.

    The joint CDF FX,Y(x,y)=P(X≀x,Y≀y)F_{X,Y}(x,y) = \mathbb{P}(X \le x, Y \le y) is the fundamental object that encodes all joint distributional information. It determines the marginals, but the marginals do not determine it.

  • 2.

    Joint PMF/PDF: For discrete RVs, PX,Y(xi,yj)=P(X=xi,Y=yj)P_{X,Y}(x_i, y_j) = \mathbb{P}(X = x_i, Y = y_j); for continuous RVs, fX,Y(x,y)=βˆ‚2FX,Y/βˆ‚xβ€‰βˆ‚yf_{X,Y}(x,y) = \partial^2 F_{X,Y} / \partial x\,\partial y. Marginals are obtained by summing or integrating out the other variable.

  • 3.

    Conditional distributions: f(y∣x)=fX,Y(x,y)/fX(x)f(y \mid x) = f_{X,Y}(x,y)/f_{X}(x). The conditional expectation E[X∣Y]\mathbb{E}[X \mid Y] satisfies the tower property E[X]=E[E[X∣Y]]\mathbb{E}[X] = \mathbb{E}[\mathbb{E}[X \mid Y]] and the law of total variance.

  • 4.

    Independence: XX and YY are independent iff fX,Y=fXβ‹…fYf_{X,Y} = f_{X} \cdot f_{Y}. Independence implies uncorrelatedness, but uncorrelated does not imply independent (except for jointly Gaussian RVs).

  • 5.

    Jacobian method: For an invertible transformation (U,V)=g(X,Y)(U,V) = g(X,Y), fU,V(u,v)=fX,Y(gβˆ’1(u,v))β‹…βˆ£Jgβˆ’1∣f_{U,V}(u,v) = f_{X,Y}(g^{-1}(u,v)) \cdot |J_{g^{-1}}|.

  • 6.

    Convolution: The PDF of Z=X+YZ = X + Y for independent X,YX, Y is fZ=fXβˆ—fYf_{Z} = f_{X} * f_{Y}. Gaussians are closed under convolution.

  • 7.

    Order statistics: FX(n)(x)=[FX(x)]nF_{X_{(n)}}(x) = [F_{X}(x)]^n for the maximum; FX(1)(x)=1βˆ’[1βˆ’FX(x)]nF_{X_{(1)}}(x) = 1 - [1 - F_{X}(x)]^n for the minimum. The minimum of i.i.d. exponentials with rate Ξ»\lambda is exponential with rate nΞ»n\lambda.

  • 8.

    Covariance and correlation: Cov(X,Y)=E[XY]βˆ’E[X]E[Y]\text{Cov}(X,Y) = \mathbb{E}[XY] - \mathbb{E}[X]\mathbb{E}[Y]; ρX,Y∈[βˆ’1,1]\rho_{X,Y} \in [-1,1]. The variance of a sum decomposes as Var(βˆ‘Xi)=βˆ‘Var(Xi)+2βˆ‘i<jCov(Xi,Xj)\text{Var}(\sum X_i) = \sum \text{Var}(X_i) + 2\sum_{i<j} \text{Cov}(X_i, X_j).

Looking Ahead

Chapter 8 extends these ideas to random vectors and the multivariate Gaussian distribution, where the covariance matrix governs everything: marginals, conditionals, and independence are all read off from the matrix. The conditional Gaussian formula β€” the Schur complement β€” is the foundation of LMMSE estimation and Kalman filtering.

πŸŽ“CommIT Contribution(2022)

Bayesian Channel Estimation via Conditional Distributions

M. Koller, B. Fesl, G. Caire β€” IEEE Transactions on Wireless Communications

Conditional distributions are the mathematical backbone of Bayesian channel estimation. Koller, Fesl, and Caire developed a scalable Bayesian MIMO channel estimator that computes the posterior fH∣Y(H∣Y)f_{\mathbf{H}|\mathbf{Y}}(\mathbf{H} | \mathbf{Y}) efficiently by exploiting the structure of the prior covariance matrix. The tower property ensures that the MMSE estimator H^=E[H∣Y]\hat{\mathbf{H}} = \mathbb{E}[\mathbf{H} | \mathbf{Y}] minimizes the average estimation error. The techniques developed in this chapter β€” conditional distributions, Bayes' rule for continuous RVs, and the conditional expectation as the MMSE estimator β€” are the theoretical foundation upon which this work rests.

channel-estimationbayesianmimoView Paper β†’