Gaussian Belief Propagation
When Messages Are Gaussians
Sum-product on discrete variables computes probability tables. For continuous variables the analog is densities β but general densities are infinite-dimensional. The algorithm becomes computationally heavy.
There is one continuous family where sum-product remains tractable: jointly Gaussian distributions. In this case, every message is itself a Gaussian density, parameterized by a mean (or precision) vector and a covariance (or precision) matrix. The message updates are finite matrix operations. This yields Gaussian belief propagation (GaBP) β the continuous analog of LDPC decoding.
The point of this section is that GaBP is the right tool whenever the model is (approximately) jointly Gaussian: iterative MIMO detection, Kalman filtering on a tree-structured state space, distributed least-squares estimation. It plugs into iterative receivers where soft outputs are needed in real time.
Definition: Gaussian Factor Graph
Gaussian Factor Graph
A Gaussian factor graph has each factor of the form where (information / precision) and (potential) are the canonical parameters. The global distribution is then the multivariate Gaussian with total precision (appropriately padded) and potential . The marginals are Gaussian with mean and variance .
The information form is the natural parameterization for message passing β products of Gaussians are computed by adding information parameters, and marginalization requires a matrix inversion that reduces nicely for tree-structured graphs.
Theorem: Gaussian BP Update Rules
For a pairwise Gaussian factor graph with factors and unary terms , the Gaussian BP messages are parameterized as with updates The final belief at node is Gaussian with precision and potential .
GaBP parameterizes messages by scalar precision and potential pairs (for pairwise factors). Products of Gaussians are Gaussians whose parameters add; marginalization of a jointly-Gaussian pair is Schur-complement computation. Every message update is , and the whole algorithm runs at matrix-inversion-level speed without ever forming an inverse.
Parameterize Gaussian messages
A Gaussian message in has the form , parameterized by . Products of such messages: .
Marginalization via Schur complement
For a pairwise factor , integrating out one variable gives a Gaussian in the other with precision and potential appropriately transformed.
Combine to get BP update
Apply the sum-product rule for the factor-to-variable message, multiply the incoming messages on the eliminated side, and marginalize. The formulas are the result.
Convergence condition
GaBP converges when the precision matrix is walk-summable, e.g., diagonally dominant. This is a widely applicable but strict condition.
Theorem: Means Are Always Correct for GaBP
If GaBP converges on a Gaussian graph with precision matrix , the computed means equal the exact Gaussian marginal means , regardless of whether the graph is a tree. The variances may differ from the true marginal variances .
This is a striking property: loopy GaBP is still correct for point estimates, even though it makes errors in uncertainty quantification. The means behave as though the graph were a tree because they solve a linear fixed-point system equivalent to the true linear system. The variances, in contrast, depend on global cycle structure that loopy BP cannot capture correctly.
Fixed-point equations are linear
Upon convergence, the GaBP potential messages satisfy a linear fixed-point system arising from the BP update rules.
Equivalence to $\mathbf{J}\boldsymbol{\mu} = \mathbf{h}$
One verifies (by direct substitution) that the computed means satisfy β precisely the system that characterizes the true Gaussian mean.
Variances miss cycle contributions
The variance estimate is based on a walk-sum expansion that includes only walks with no backtracking; actual includes all walks. Cycles contribute discounted corrections that GaBP misses.
Gaussian Belief Propagation (Pairwise)
Complexity: per iteration (one scalar update per edge). Total memory: . Compare to direct matrix inversion β GaBP is exponentially cheaper for sparse .Walk-summable (diagonally dominant) guarantees convergence. In practice, damping messages (take a convex combination of old and new) extends the convergence range.
Example: GaBP on a 3-Node Cycle
Solve the linear system with and using GaBP. Compare to the exact solution.
Exact solution
, so .
Set up GaBP
Pairwise factors: for ; diagonal . All three pairwise edges. The graph is a 3-cycle.
Run iterations
Initialize . Iterate until convergence. After ~20 iterations, β exactly matches the true means, despite the cycle.
Variances differ
True variances: . GaBP variance estimate: β slightly off. This illustrates the theorem: means exact, variances approximate.
Why This Matters: GaBP for MIMO Detection
For a MIMO receiver with dense channel matrix , exact MMSE detection requires inverting β cubic in the number of streams. GaBP on the factor graph of the posterior gives MMSE estimates (means) with iteration complexity linear in 's non-zeros. For sparse or banded channels (ISI, massive MIMO with sparse beamspace representation), this delivers orders of magnitude speedup. 5G terminal MIMO receivers often use a GaBP-based inner loop coupled with an LDPC outer decoder.
GaBP Convergence: Diagonal Dominance Matters
Show GaBP convergence on random sparse Gaussian graphs. Vary the diagonal dominance factor and observe convergence vs. divergence.
Parameters
Common Mistake: GaBP Divergence from Weak Diagonal Dominance
Mistake:
Running GaBP on a precision matrix with (fails strict diagonal dominance) and assuming it will converge.
Correction:
GaBP is guaranteed to converge only when is walk-summable (implied by strict diagonal dominance). For weakly diagonally dominant or indefinite matrices, add damping: , with . In adversarial cases (e.g., MIMO with coherent channels), fall back to direct solvers.
GaBP in Silicon
GaBP's scalar message updates map cleanly to parallel hardware: each edge's update is independent within a layer. Commercial 5G receivers implement GaBP with damping factor 0.5 and 5-10 iterations, achieving within 0.5 dB of exact MMSE at a fraction of the matrix-inversion latency.
- β’
Precision required: ~12 bits to avoid variance drift.
- β’
Walk-summability checked offline; online diagnostic monitors divergence.
- β’
Typical convergence in 5-15 iterations for MIMO detection.
Approximate Gaussian BP for Massive MIMO
The CommIT group and collaborators have developed GaBP-based MIMO detection algorithms tailored for massive MIMO uplink. These algorithms combine GaBP with deterministic-equivalent regularization (Chapter 16) to achieve near-MMSE performance with hardware-friendly complexity.
Quick Check
GaBP is run on a loopy Gaussian graph and converges. What can we say about the computed means and variances?
Means exact, variances approximate
Both means and variances exact
Both means and variances approximate
Neither (GaBP cannot handle loops)
This is the key theorem: the means solve exactly, but variances miss cycle contributions.
Key Takeaway
Gaussian belief propagation is sum-product specialized to jointly Gaussian distributions. Messages are Gaussian densities parameterized by information form ; products add, marginalization is Schur-complement. On any convergent graph (including loopy ones), the means are exact β variances are approximate. GaBP is the workhorse for iterative MIMO detection and large-scale sparse linear system solving.