Generalized Linear Models

Beyond Additive Gaussian Noise

The standard RF imaging model y=Ac+w\mathbf{y} = \mathbf{A}\mathbf{c} + \mathbf{w} with w∼CN(0,Οƒ2I)\mathbf{w} \sim \mathcal{CN}(\mathbf{0}, \sigma^2\mathbf{I}) is a special case of a much broader class: Generalized Linear Models (GLMs).

Three RF-relevant scenarios require non-Gaussian output channels:

  • 1-bit receivers: Low-cost hardware quantizes each measurement to a single bit: ym=sign(β„œ{zm})y_m = \text{sign}(\Re\{z_m\}). The measurement discards all amplitude information β€” only the sign survives.
  • Power-only / envelope detection: ym=∣zm∣2+noisey_m = |z_m|^2 + \text{noise}. Used in passive coherent location and angle-of-arrival systems.
  • Photon-counting / shot noise: ym∼Poisson(Ξ»m)y_m \sim \text{Poisson}(\lambda_m) where Ξ»m=∣zm∣2\lambda_m = |z_m|^2. Relevant for optical/THz coherent sensing.

GAMP handles all these via a single interface: the output function gout(ym,p^m,Ο„p)g_{\text{out}}(y_m, \hat{p}_m, \tau_p) replaces the Gaussian residual computation and encapsulates the non-standard likelihood.

Definition:

Generalized Linear Model (GLM)

A generalized linear model (GLM) for RF imaging has the form:

ym∼p(ym∣zm),zm=amTc,xq∼p0(xq),y_m \sim p(y_m \mid z_m), \quad z_m = \mathbf{a}_m^T \mathbf{c}, \quad x_q \sim p_0(x_q),

where:

  • p(ym∣zm)p(y_m \mid z_m) is the output channel (likelihood), a function of the linear mixing output zmz_m alone (conditional independence across mm).
  • p0(xq)p_0(x_q) is the input channel (prior) on each scene element.
  • The posterior is p(c∣y)∝∏qp0(xq)β‹…βˆmp(ym∣amTc)p(\mathbf{c} \mid \mathbf{y}) \propto \prod_q p_0(x_q) \cdot \prod_m p(y_m \mid \mathbf{a}_m^T \mathbf{c}).

The Gaussian case p(y∣z)=CN(y;z,Οƒ2)p(y \mid z) = \mathcal{CN}(y; z, \sigma^2) is the standard CS model. GAMP applies to any GLM where the output channel admits efficient computation of the posterior mean and variance.

Definition:

GAMP Output Function goutg_{\text{out}}

For any output channel p(y∣z)p(y \mid z), define the output function:

gout(y,p^,Ο„p)β‰œz^postβˆ’p^Ο„p,Ο„sβ‰œβˆ’βˆ‚goutβˆ‚p^,g_{\text{out}}(y, \hat{p}, \tau_p) \triangleq \frac{\hat{z}_{\text{post}} - \hat{p}}{\tau_p}, \quad \tau_s \triangleq -\frac{\partial g_{\text{out}}}{\partial \hat{p}},

where z^post=E[z∣y,z∼N(p^,Ο„p)]\hat{z}_{\text{post}} = \mathbb{E}[z \mid y, z \sim \mathcal{N}(\hat{p}, \tau_p)] is the posterior mean of zz under the combined channel:

p(z∣y,p^,Ο„p)∝p(y∣z) N(z;p^,Ο„p).p(z \mid y, \hat{p}, \tau_p) \propto p(y \mid z)\,\mathcal{N}(z; \hat{p}, \tau_p).

Interpretation: goutg_{\text{out}} computes the MMSE estimate of zz given the measurement ymy_m and a Gaussian "cavity" distribution N(p^,Ο„p)\mathcal{N}(\hat{p}, \tau_p) representing the extrinsic belief on zz from all other variables. The ratio (z^postβˆ’p^)/Ο„p(\hat{z}_{\text{post}} - \hat{p})/\tau_p is the normalized innovation.

Example: Output Function for Gaussian Likelihood

Derive gout(y,p^,Ο„p)g_{\text{out}}(y, \hat{p}, \tau_p) for p(y∣z)=N(y;z,Οƒ2)p(y \mid z) = \mathcal{N}(y; z, \sigma^2). Verify that GAMP reduces to standard AMP in this case.

Example: Output Function for 1-Bit (Probit) Likelihood

Derive goutg_{\text{out}} for the 1-bit measurement model: p(y∣z)=Ξ¦(y z/Οƒdither)p(y \mid z) = \Phi(y\,z/\sigma_{\text{dither}}), where y∈{+1,βˆ’1}y \in \{+1, -1\} and Ξ¦(β‹…)\Phi(\cdot) is the Gaussian CDF (probit link).

,

Output Channel

In a generalized linear model (GLM), the output channel p(ym∣zm)p(y_m \mid z_m) specifies the conditional distribution of the measurement ymy_m given the linear mixing output zm=amTxz_m = \mathbf{a}_m^T\mathbf{x}. In standard compressed sensing, the output channel is Gaussian. GAMP can handle any output channel for which the posterior mean E[z∣y,z∼N(p^,Ο„p)]\mathbb{E}[z \mid y, z \sim \mathcal{N}(\hat{p}, \tau_p)] is computable.

Related: Generalized Linear Model (GLM), Output Function

Mills Ratio

The Mills ratio (or hazard function of the Gaussian) is Ο•(x)/Ξ¦(x)\phi(x)/\Phi(x), where Ο•\phi is the standard Gaussian PDF and Ξ¦\Phi is the CDF. It appears in the GAMP output function for probit (1-bit) and truncated-Gaussian likelihoods. Numerically, it is computed as 2 erfcx(x/2) eβˆ’x2/2/2\sqrt{2}\,\text{erfcx}(x/\sqrt{2})\,e^{-x^2/2}/2 for stability in the tail region x≫1x \gg 1.

Related: Output Channel, 1 Bit Compressed Sensing

GAMP Output Channels for RF Imaging

Likelihood ModelExpression p(y∣z)p(y|z)goutg_{\text{out}} Key FormulaRF Application
GaussianN(y;z,Οƒ2)\mathcal{N}(y; z, \sigma^2)(yβˆ’p^)/(Οƒ2+Ο„p)(y - \hat{p})/(\sigma^2 + \tau_p)Standard CS, homodyne radar
1-bit (probit)Ξ¦(yz/Οƒ)\Phi(yz/\sigma), y∈{+1,βˆ’1}y\in\{+1,-1\}Mills ratio (inverse): yΟ•(u)/Ξ¦(yu)y\phi(u)/\Phi(yu)Low-cost ADC, compressive receivers
Power-onlyp(y∣z)∝eβˆ’y/∣z∣2/∣z∣2p(y|z) \propto e^{-y/|z|^2}/|z|^2Truncated Gaussian posteriorPhase-less radar, passive sensing
Poissoneyzβˆ’ez/y!e^{yz - e^z}/y!Laplace approx: yβˆ’ep^+Ο„p/2y - e^{\hat{p}+\tau_p/2}Photon-counting, THz sensing
Logistic (binary)Οƒ(yz)=1/(1+eβˆ’yz)\sigma(yz) = 1/(1+e^{-yz})Probit approximationOccupancy grid mapping

Theorem: State Evolution for GAMP

Under i.i.d. Gaussian A\mathbf{A} with i.i.d. entries CN(0,1/M)\mathcal{CN}(0, 1/M), as M,Nβ†’βˆžM, N \to \infty with M/Nβ†’Ξ΄M/N \to \delta, the GAMP state variables satisfy:

Ο„pt=Ο„xtΞ΄,Ο„st=EY,Z,W ⁣[βˆ’βˆ‚goutβˆ‚p^(Y,Z+Ο„pt W,Ο„pt)],\tau_p^t = \frac{\tau_x^t}{\delta}, \quad \tau_s^t = \mathbb{E}_{Y,Z,W}\!\left[-\frac{\partial g_{\text{out}}}{\partial\hat{p}}(Y, Z + \sqrt{\tau_p^t}\,W, \tau_p^t)\right],

(Ο„rt+1)βˆ’1=δ τst,Ο„xt+1=EX0,Z ⁣[(gin ⁣(X0+Ο„rt+1 Z,Ο„rt+1)βˆ’X0)2],(\tau_r^{t+1})^{-1} = \delta\,\tau_s^t, \quad \tau_x^{t+1} = \mathbb{E}_{X_0, Z}\!\left[\left(g_{\text{in}}\!\left(X_0 + \sqrt{\tau_r^{t+1}}\,Z, \tau_r^{t+1}\right) - X_0\right)^2\right],

where expectations are over the joint distributions (Y,Z)∼p0Γ—p(β‹…βˆ£z)(Y, Z) \sim p_0 \times p(\cdot | z) and W,Z∼CN(0,1)W, Z \sim \mathcal{CN}(0,1).

State evolution predicts the GAMP MSE without running the algorithm: it reduces the high-dimensional problem to a pair of scalar recursions. For non-Gaussian likelihoods, the effective noise seen by the input denoiser is determined by the average Jacobian of goutg_{\text{out}} β€” a scalar quantity that captures the "information content" of one measurement.

Why This Matters: 1-Bit CS for Low-Cost RF Imaging Receivers

High-resolution ADCs (12–16 bits) consume significant power: roughly 1 mW/GHz per bit of resolution. For a wideband radar with W=1W = 1 GHz, a 12-bit ADC at Nyquist rate consumes ∼12\sim 12 mW β€” per receiver channel.

1-bit receivers replace each ADC with a single comparator (< 0.1 mW), enabling dense receiver arrays at a fraction of the power. The GAMP output function for the probit model recovers near-standard performance at oversampling ratios M/Nβ‰₯2M/N \geq 2–33, with a penalty of ∼2\sim 2 dB compared to full-precision measurements.

This is the core design principle behind the CommIT group's compressive RF imaging receivers in Chapter 11.

See full treatment in Multi-View, Multi-Frequency Sensing Geometry

1-Bit CS: GAMP vs Mismatched Gaussian Approximation

Reconstruction NMSE as a function of oversampling ratio M/NM/N for the 1-bit measurement model. GAMP (correct model) uses the probit output function; Mismatched GAMP treats the signed measurements as linear Gaussian observations.

At high oversampling (M/N>3M/N > 3), both approaches converge; at low oversampling (M/N<1.5M/N < 1.5), the model-matched GAMP gives 5–10 dB advantage.

Parameters
0.1
20

GAMP Convergence for Different Likelihood Models

NMSE versus iteration for three likelihood models: Gaussian (standard CS), 1-bit, and Poisson. The solid curve uses the correct GAMP output function; the dashed curve shows mismatched GAMP (always uses Gaussian likelihood).

At convergence, the model-matched GAMP reaches the Bayes-optimal MSE predicted by state evolution; mismatched GAMP converges to a higher floor.

Parameters
1.5

Common Mistake: GAMP Convergence Is Fragile for Non-Log-Concave Likelihoods

Mistake:

GAMP's convergence guarantees (via state evolution) require the output channel to be log-concave in zz β€” i.e., βˆ’log⁑p(y∣z)-\log p(y|z) is convex. For Poisson, power-only, and logistic channels this fails for certain measurement regimes, creating multiple Bethe free energy minima.

Symptoms: oscillation in the residual, divergence of Ο„st\tau_s^t, or NMSE stuck well above the SE prediction.

Correction:

(1) Damp the GAMP updates: s^mt+1←αgout+(1βˆ’Ξ±)s^mt\hat{s}_m^{t+1} \leftarrow \alpha g_{\text{out}} + (1-\alpha)\hat{s}_m^t with Ξ±=0.5\alpha = 0.5–0.80.8. (2) For structured sensing matrices (real-world physical arrays, not i.i.d. Gaussian), switch to the VAMP framework (Chapter 18) which has more robust convergence properties. (3) Monitor Ο„st\tau_s^t: if it becomes negative or exceeds 1/Ο„p1/\tau_p, reset to the Gaussian output function temporarily.

πŸ”§Engineering Note

1-Bit ADC Implementation Constraints

Practical 1-bit compressed sensing receivers use a comparator with a programmable threshold (dithering) to break the symmetry of the sign measurement. Without dithering, signed measurements cannot distinguish between xx and 2x2x, making amplitude recovery impossible.

The dither signal is typically a pseudo-random sequence with known statistical properties, which enters the probit model as Οƒdither\sigma_{\text{dither}}. Optimal dither variance is Οƒdither2β‰ˆ0.5 Var[zm]\sigma_{\text{dither}}^2 \approx 0.5\,\text{Var}[z_m], striking a balance between information preservation and quantization distortion.

Practical Constraints
  • β€’

    Comparator latency: 1–5 ps at GHz rates (vs. 50–100 ps for 12-bit ADC)

  • β€’

    Power: < 1 mW/GHz per comparator vs. ~12 mW/GHz for 12-bit ADC

  • β€’

    Dither must be synchronized across all receiver channels for coherent processing

  • β€’

    GAMP with probit model requires M/Nβ‰₯2M/N \geq 2 for reliable recovery at SNR = 20 dB

Historical Note: 1-Bit Compressed Sensing: Origins

2008–2015

The 1-bit compressed sensing problem was introduced by Boufounos and Baraniuk in 2008 (IEEE ICASSP), who showed that sparse signals are recoverable from their sign measurements alone. The key insight was that the sign function preserves the geometry of sparse vectors in a metric space sense.

Optimal Bayesian recovery via GAMP for 1-bit measurements was developed by Schniter and Rangan (2015, IEEE Trans. SP), who derived the probit output function and showed that GAMP achieves near-oracle MSE even for M/N<1M/N < 1 when the signal is sparse enough.

Quick Check

For Gaussian likelihood p(y∣z)=N(y;z,Οƒ2)p(y|z) = \mathcal{N}(y; z, \sigma^2), the GAMP output function gout(y,p^,Ο„p)g_{\text{out}}(y, \hat{p}, \tau_p) evaluates to:

(yβˆ’p^)/Ο„p(y - \hat{p})/\tau_p

(yβˆ’p^)/(Οƒ2+Ο„p)(y - \hat{p})/(\sigma^2 + \tau_p)

(yβˆ’p^)/Οƒ2(y - \hat{p})/\sigma^2

Ξ¦((yβˆ’p^)/Οƒ2+Ο„p)\Phi((y - \hat{p})/\sqrt{\sigma^2+\tau_p})

Key Takeaway

GLMs extend the CS model by replacing the Gaussian noise assumption with any separable output channel p(ym∣zm)p(y_m | z_m). GAMP handles GLMs via a single interface: the output function goutg_{\text{out}}, which computes the posterior mean of zmz_m under the combined cavity + likelihood distribution. For 1-bit receivers, the probit goutg_{\text{out}} (Mills ratio) recovers near-standard performance at 2–3Γ— oversampling β€” enabling dense, low-power receiver arrays in RF imaging systems.