Joint VR and Channel Estimation

Why Estimate Both at Once

Sections 18.1–18.4 have given us three separate components: a VR mask with a 2D Markov prior (18.2), a subarray processing pipeline (18.3), and a near-field sparse channel representation (18.4). Running them sequentially β€” first detect the VR, then estimate the channel β€” is tempting but wastes information. A sequential scheme uses the raw pilot correlator for VR detection and throws away the channel coefficients; a proper scheme feeds the channel estimate back into the VR detector because a large coherent signal on an antenna is stronger evidence of mk,n=1m_{k,n} = 1 than a raw energy test. The principled machinery for this feedback is the EM algorithm with the 2D Markov prior playing the role of a structured latent variable and the near-field dictionary supplying the observation model.

,

Definition:

Joint Posterior of Mask and Channel

Stack all pilot observations into Yp∈CNtΓ—Ο„p\mathbf{Y}_p \in \mathbb{C}^{N_t \times \tau_p}. For user kk, let mk\mathbf{m}_k denote the latent binary mask and Hk=Apolarzk\mathbf{H}_{k} = \mathbf{A}_{\text{polar}} \mathbf{z}_k the near-field channel in polar coordinates. Under the Gaussian observation model and with the 2D Markov prior on mk\mathbf{m}_k (Definition D2D Markov Random Field Prior on the VR Mask), the joint posterior factors as Pr⁑[mk,zk∣Yp]∝Pr⁑[mk]⏟2DΒ Isingβ‹…Pr⁑[zk]⏟sparsityβ‹…p(Yp∣mk,zk)⏟Gaussian.\Pr[\mathbf{m}_k, \mathbf{z}_k \mid \mathbf{Y}_p] \propto \underbrace{\Pr[\mathbf{m}_k]}_{\text{2D Ising}} \cdot \underbrace{\Pr[\mathbf{z}_k]}_{\text{sparsity}} \cdot \underbrace{p(\mathbf{Y}_p \mid \mathbf{m}_k, \mathbf{z}_k)}_{\text{Gaussian}}. The Gaussian likelihood is p(Yp∣mk,zk)=CN(Yp; D(mk)Apolarzk Si,kk, σ2I),p(\mathbf{Y}_p \mid \mathbf{m}_k, \mathbf{z}_k) = \mathcal{CN}\bigl(\mathbf{Y}_p;\, \mathbf{D}(\mathbf{m}_k) \mathbf{A}_{\text{polar}} \mathbf{z}_k\, {\mathbf{S}_{i,k}}_{k},\, \sigma^2 \mathbf{I}\bigr), where D(mk)=diag(mk)\mathbf{D}(\mathbf{m}_k) = \text{diag}(\mathbf{m}_k) applies the mask element-wise.

The log-joint L(mk,zk)=log⁑Pr⁑[mk,zk∣Yp]\mathcal{L}(\mathbf{m}_k, \mathbf{z}_k) = \log \Pr[\mathbf{m}_k, \mathbf{z}_k \mid \mathbf{Y}_p] is a sum of a discrete-MRF term in mk\mathbf{m}_k, a continuous quadratic term in zk\mathbf{z}_k, and a bilinear coupling mkTβ‹…\mathbf{m}_k^T \cdot (quadratic in zk\mathbf{z}_k). Maximizing jointly over (mk,zk)(\mathbf{m}_k, \mathbf{z}_k) is NP-hard in general, but alternating maximization (EM) converges to a good local optimum in a few iterations.

Theorem: Monotone Ascent of EM for the Joint Problem

Let q(t)(mk)q^{(t)}(\mathbf{m}_k) denote the variational distribution over mk\mathbf{m}_k at EM iteration tt and z^k(t)\hat{\mathbf{z}}_k^{(t)} the channel estimate. Define the evidence lower bound ELBO(q,z)=Eq ⁣[log⁑Pr⁑[mk]]+Eq ⁣[log⁑p(Yp∣mk,z)]+H(q)+log⁑Pr⁑[z].\text{ELBO}(q, \mathbf{z}) = \mathbb{E}_{q}\!\left[\log \Pr[\mathbf{m}_k]\right] + \mathbb{E}_{q}\!\left[\log p(\mathbf{Y}_p \mid \mathbf{m}_k, \mathbf{z})\right] + H(q) + \log \Pr[\mathbf{z}]. The EM updates β€” E-step: q(t+1)=arg⁑max⁑qELBO(q,z^k(t))q^{(t+1)} = \arg\max_q \text{ELBO}(q, \hat{\mathbf{z}}_k^{(t)}) subject to the mean-field factorization q(mk)=∏nqn(mk,n)q(\mathbf{m}_k) = \prod_n q_n(m_{k,n}), and M-step: z^k(t+1)=arg⁑max⁑zELBO(q(t+1),z)\hat{\mathbf{z}}_k^{(t+1)} = \arg\max_{\mathbf{z}} \text{ELBO}(q^{(t+1)}, \mathbf{z}) β€” produce a sequence {ELBO(t)}\{\text{ELBO}^{(t)}\} that is monotonically non-decreasing.

Each step maximizes the ELBO with respect to a different argument, so the ELBO cannot decrease. Because the ELBO is bounded above by the log-evidence log⁑p(Yp)\log p(\mathbf{Y}_p), the sequence converges. The limit is a stationary point β€” not necessarily the global maximum β€” but in practice the 2D Markov prior regularizes the landscape well and good initializations reach high-quality solutions within 3–5 iterations.

,

Joint VR + Channel Estimation via EM

Complexity: O(Tβ‹…(KBPNt+LGNt))\mathcal{O}(T \cdot (K_{\text{BP}} N_t + L G N_t)). For T=5T = 5 EM outer iterations, KBP=10K_{\text{BP}} = 10 BP sweeps, L=4L = 4, and G=10NtG = 10N_t, total ∼5β‹…(10Nt+40Nt2)\sim 5 \cdot (10 N_t + 40 N_t^{2}) flops. Well within the subarray complexity budget of Section 18.3.
Input: Pilot observations Yp\mathbf{Y}_p, pilot Si,kk{\mathbf{S}_{i,k}}_{k}, polar
dictionary Apolar\mathbf{A}_{\text{polar}}, MRF parameters (J,h)(J, h), noise
variance Οƒ2\sigma^2, maximum iterations TT, convergence tolerance
Ο΅\epsilon.
Output: VR marginals {qn}\{q_n\} and channel estimate H^k\hat{\mathbf{H}}_k.
1. Initialize: set qn(0)←σ(β„“n)q_n^{(0)} \leftarrow \sigma(\ell_n) from the raw
per-antenna LLR of Definition DPosterior MRF from Pilot Observations; set
z^k(0)\hat{\mathbf{z}}_k^{(0)} from polar-OMP (Algorithm APolar-OMP for Near-Field Channel Estimation)
on the full-aperture pilot.
2. for t=0,1,…,Tβˆ’1t = 0, 1, \ldots, T-1 do
3. \quad E-step (Loopy BP on the MRF):
4. \quad\quad Compute data-driven LLRs
β„“n(t)=log⁑p(yp,n∣mk,n=1,z^k(t))βˆ’log⁑p(yp,n∣mk,n=0)\ell_n^{(t)} = \log p(\mathbf{y}_{p,n} \mid m_{k,n} = 1, \hat{\mathbf{z}}_k^{(t)}) - \log p(\mathbf{y}_{p,n} \mid m_{k,n} = 0).
5. \quad\quad Run KBPK_{\text{BP}} sweeps of sum-product BP on the 2D Ising
graph with external fields hn=h+12β„“n(t)h_n = h + \tfrac{1}{2}\ell_n^{(t)}.
6. \quad\quad Extract marginals qn(t+1)=Pr⁑[mk,n=1∣Yp,z^k(t)]q_n^{(t+1)} = \Pr[m_{k,n} = 1 \mid \mathbf{Y}_p, \hat{\mathbf{z}}_k^{(t)}].
7. \quad M-step (weighted sparse LS):
8. \quad\quad Form weight matrix W(t+1)=diag(qn(t+1))\mathbf{W}^{(t+1)} = \text{diag}(q_n^{(t+1)}) and solve
z^k(t+1)=arg⁑min⁑zβˆ₯W(t+1)(YpSi,kkβˆ—/βˆ₯Si,kkβˆ₯2βˆ’Apolarz)βˆ₯22+Ξ»βˆ₯zβˆ₯1\hat{\mathbf{z}}_k^{(t+1)} = \arg\min_\mathbf{z} \| \mathbf{W}^{(t+1)} (\mathbf{Y}_p {\mathbf{S}_{i,k}}_{k}^* /\|{\mathbf{S}_{i,k}}_{k}\|^2 - \mathbf{A}_{\text{polar}} \mathbf{z}) \|_2^2 + \lambda \|\mathbf{z}\|_1.
9. \quad Check convergence: if
βˆ₯z^k(t+1)βˆ’z^k(t)βˆ₯2<Ο΅\|\hat{\mathbf{z}}_k^{(t+1)} - \hat{\mathbf{z}}_k^{(t)}\|_2 < \epsilon
break.
10. end for
11. Assemble H^k←diag(q(T)) Apolar z^k(T)\hat{\mathbf{H}}_k \leftarrow \text{diag}(\mathbf{q}^{(T)})\, \mathbf{A}_{\text{polar}}\, \hat{\mathbf{z}}_k^{(T)}.
12. return {qn(T)},H^k\{q_n^{(T)}\}, \hat{\mathbf{H}}_k.

Step 5 is the core CommIT contribution: loopy BP on the 2D Ising graph enforces spatial smoothness of the VR and acts as a structured regularizer on the mask. Without it, the mean-field update would amount to independent sigmoid thresholding per antenna β€” the same as sequential VR detection, and strictly worse than the joint scheme (Theorem TMonotone Ascent of EM for the Joint Problem still holds but the operating point is worse).

Joint EM vs Sequential Estimation: NMSE vs VR Mismatch

Compare four estimators across the operating SNR: (i) genie (true VR + MMSE), (ii) sequential (hard-threshold VR detector + MMSE on detected support), (iii) Xu–Caire joint EM (this section), (iv) LS on the full aperture. At moderate SNR the joint EM lies within 1 dB of the genie; the sequential detector suffers a 4–6 dB penalty near the boundary of VR ambiguity.

Parameters
32
32
0.3
16
-5
20

Example: How Many EM Iterations Do We Need?

On a 32Γ—3232 \times 32 panel with VRΒ fraction=0.3\text{VR fraction} = 0.3 and SNR=0\text{SNR} = 0 dB, how does the NMSE of the joint EM estimator evolve with EM iteration index tt and what is a reasonable stopping rule?

Pilot Overhead in the XL-MIMO Regime

A subtle payoff of the joint estimator is that it shrinks the pilot overhead needed to achieve a target NMSE. With the Markov prior exploited, pilot lengths as short as Ο„p=8\tau_p = 8 symbols suffice on a 32Γ—3232 \times 32 panel at 1010 dB SNR β€” less than the K=16K = 16 suggested by orthogonal pilot allocation. The remaining non-orthogonality is absorbed by the MMSE / sparse step, which the 2D prior makes robust. In short: the 2D Markov prior partially substitutes for pilot orthogonality, a surprisingly aggressive form of pilot decontamination specific to XL-MIMO.

Common Mistake: Do Not Freeze the Mask Too Early

Mistake:

After the first BP pass, the marginals qnq_n look well-separated, so round them to a hard 0/1 mask and finish with an ordinary LS on the hard support.

Correction:

Hard-thresholding between EM iterations destroys the soft evidence that the M-step needs to refine the channel. The penalty is largest at the boundary of the VR where qnq_n sits around 0.3–0.7, precisely the antennas where the joint information flow matters most. Keep the marginals soft throughout EM and hard-threshold only at the very end, and only if your downstream combiner requires a binary mask. The interactive plot above shows the 2–4 dB NMSE penalty of premature thresholding at moderate SNR.

⚠️Engineering Note

Deploying the Joint Estimator in Real Time

A production XL-MIMO deployment running the joint EM estimator of Algorithm AJoint VR + Channel Estimation via EM must balance three constraints: coherence block length (typically Ο„c=100\tau_c = 100–200200 symbols), per-user baseband budget (typically <1< 1 ms on embedded DSP), and fronthaul capacity. A practical blueprint:

  • Outer loop T=3T = 3–55. More iterations rarely improve NMSE by more than 0.10.1 dB.
  • Inner BP KBP=10K_{\text{BP}} = 10 sweeps. Use a checkerboard schedule so the sweeps parallelize across half the grid at a time.
  • Polar dictionary cached. The dictionary depends only on array geometry and wavelength, not on channel or user state, so it can be precomputed once and reused for months.
  • Subarray fallback. If the per-user budget is tight, run the joint EM on the active subarrays only (Section 18.3), which drops the M-step cost by S/∣Ak∣S / |\mathcal{A}_k| without hurting NMSE noticeably.
  • Graceful degradation. When SNR drops below βˆ’5-5 dB, fall back to polar- OMP with a flat mask prior; the MRF benefit vanishes under very noisy evidence and the BP sweeps waste cycles.
Practical Constraints
  • β€’

    Outer EM iterations T∈[3,5]T \in [3, 5]

  • β€’

    Inner BP sweeps KBP∈[8,16]K_{\text{BP}} \in [8, 16]

  • β€’

    Per-user runtime budget: <0.5< 0.5 ms on embedded DSP for Nt≀4096N_t \leq 4096

  • β€’

    Fallback threshold: SNR <βˆ’5< -5 dB reverts to polar-OMP with uniform mask

,