Sums of Independent Random Variables and Convolution

Convolution: The Hard Way and the Easy Way

Finding the distribution of Z=X+YZ = X + Y for independent continuous RVs requires the convolution integral: fZ(z)=βˆ«βˆ’βˆžβˆžfX(x) fY(zβˆ’x) dxf_{Z}(z) = \int_{-\infty}^{\infty}f_{X}(x)\,f_{Y}(z-x)\,dx. For a sum of nn independent RVs, this becomes an (nβˆ’1)(n-1)-fold iterated integral β€” increasingly unmanageable.

Transforms convert convolution to multiplication: Ο•XZ=Ο•XXβ‹…Ο•XY{\phi_X}_{Z} = {\phi_X}_{X} \cdot {\phi_X}_{Y}. Multiply, then invert. This is the fundamental strategy of this section.

Theorem: The Convolution Theorem

Let XX and YY be independent random variables. The PDF of Z=X+YZ = X + Y is the convolution fZ=fXβˆ—fYf_{Z} = f_{X} * f_{Y}, and the CF satisfies

Ο•Z(u)=Ο•X(u)β‹…Ο•Y(u).\phi_Z(u) = \phi_X(u) \cdot \phi_Y(u).

Conversely, if Ο•Z(u)=Ο•X(u)Ο•Y(u)\phi_Z(u) = \phi_X(u)\phi_Y(u) and Ο•X(u)β‰ 0\phi_X(u) \neq 0 in a neighborhood of 00, then Ο•Y(u)=Ο•Z(u)/Ο•X(u)\phi_Y(u) = \phi_Z(u)/\phi_X(u) β€” deconvolution in the transform domain.

Convolution in the "time" (probability) domain corresponds to multiplication in the "frequency" (transform) domain. This is the same principle that makes Fourier analysis powerful in signal processing.

Example: Sum of Independent Gaussian Random Variables

Let Xi∼N(ΞΌi,Οƒi2)X_i \sim \mathcal{N}(\mu_i, \sigma_i^2) be independent for i=1,…,ni = 1, \ldots, n. Find the distribution of Sn=βˆ‘i=1nXiS_n = \sum_{i=1}^n X_i.

Key Takeaway

The Gaussian distribution is stable under addition: a sum of independent Gaussians is Gaussian. This is the reason the Gaussian plays such a central role β€” it is the only distribution (up to trivial scaling) that is a fixed point of the operation "sum, then rescale," which is exactly what the CLT describes.

Example: Sum of Independent Exponentials is Gamma

Let X1,…,XnX_1, \ldots, X_n be i.i.d. Exp(Ξ»)\text{Exp}(\lambda). Find the distribution of Sn=βˆ‘i=1nXiS_n = \sum_{i=1}^n X_i.

Convolution via Characteristic Functions

This plot shows two independent distributions XX and YY (top), their characteristic functions (middle), and the distribution of Z=X+YZ = X + Y obtained by multiplying the CFs and inverting (bottom). Adjust the distributions to see how the transform domain simplifies convolution.

Parameters
1
1
πŸ”§Engineering Note

Thermal Noise as a Sum of Independent Contributions

In electronic systems, thermal noise arises from the random motion of electrons in resistive elements. The total noise voltage is a sum of contributions from many independent sources β€” different circuit components, amplifier stages, and antenna elements. By the CLT (Section Β§The Law of Large Numbers and Central Limit Theorem), this sum is approximately Gaussian regardless of the distribution of individual contributions. The transform approach makes this rigorous: the CF of the sum converges to a Gaussian CF as the number of terms grows.

This is the physical justification for the AWGN (additive white Gaussian noise) model that pervades communications theory.

Practical Constraints
  • β€’

    Requires many independent contributions of comparable magnitude

  • β€’

    Fails when one dominant interferer is present (impulsive noise)

Common Mistake: Convolution Requires Independence

Mistake:

Writing Ο•X+Y(u)=Ο•X(u)Ο•Y(u)\phi_{X+Y}(u) = \phi_X(u)\phi_Y(u) without verifying that XX and YY are independent. If XX and YY are dependent, the CF of the sum is Ο•X+Y(u)=E[eju(X+Y)]\phi_{X+Y}(u) = \mathbb{E}[e^{ju(X+Y)}], which does not factor.

Correction:

The product rule applies only when XβŠ₯YX \perp Y. For dependent random variables, you must work with the joint CF Ο•X,Y(u,v)\phi_{X,Y}(u, v) and set v=uv = u to get Ο•X+Y(u)\phi_{X+Y}(u).