Fronthaul Requirements and Compression

The Fronthaul Bottleneck

Every cooperation level requires some information exchange between the APs and the CPU. Even Level 1 must forward ∣Dm∣|\mathcal{D}_m| complex scalars per channel use. In practice, these complex numbers must be quantized to a finite number of bits for digital transmission over the fronthaul link. The quantization introduces distortion that degrades the effective SINR. The system designer faces a tradeoff: more quantization bits reduce distortion but increase the fronthaul rate requirement. This section develops the theory of fronthaul-aware distributed processing: how many bits are needed, how to allocate them across APs and users, and what happens when the fronthaul capacity is the binding constraint.

Definition:

Scalar Quantization of Local Estimates

In distributed processing (Levels 1–3), AP mm computes local estimates s^mk∈C\hat{s}_{mk} \in \mathbb{C} for each k∈Dmk \in \mathcal{D}_m and quantizes them before fronthaul transmission. Using a bb-bit uniform scalar quantizer on the real and imaginary parts independently, the quantized estimate is

s^mk(q)=s^mk+emk\hat{s}_{mk}^{(q)} = \hat{s}_{mk} + e_{mk}

where emke_{mk} is the quantization error with variance

Οƒq2=Ξ”212=(s^maxβ‘βˆ’s^min⁑)212β‹…22b\sigma^2_{q} = \frac{\Delta^2}{12} = \frac{(\hat{s}_{\max} - \hat{s}_{\min})^2}{12 \cdot 2^{2b}}

per real dimension, where Ξ”=(s^maxβ‘βˆ’s^min⁑)/2b\Delta = (\hat{s}_{\max} - \hat{s}_{\min}) / 2^b is the quantization step size. The fronthaul rate per AP is

Rfh,m=2b∣Dm∣fs[bit/s]R_{\text{fh},m} = 2b |\mathcal{D}_m| f_s \quad [\text{bit/s}]

where fsf_s is the sampling rate (equal to the system bandwidth for Nyquist sampling).

The factor of 2 accounts for quantizing both real and imaginary parts. With b=8b = 8 bits, ∣Dm∣=10|\mathcal{D}_m| = 10 users, and fs=20f_s = 20 MHz: Rfh,m=2Γ—8Γ—10Γ—20Γ—106=3.2R_{\text{fh},m} = 2 \times 8 \times 10 \times 20 \times 10^6 = 3.2 Gbit/s per AP. This is within the capacity of a single 10G Ethernet fronthaul link.

Definition:

Fronthaul Capacity Constraint

The fronthaul link between AP mm and the CPU has a finite capacity CfhC_{\text{fh}} (in bit/s). The quantization resolution is constrained by

2bm∣Dm∣fs≀Cfh2 b_m |\mathcal{D}_m| f_s \leq C_{\text{fh}}

which gives the maximum number of quantization bits per AP:

bm≀Cfh2∣Dm∣fsb_m \leq \frac{C_{\text{fh}}}{2 |\mathcal{D}_m| f_s}

For Level 4 (centralized processing), the fronthaul must carry the full received signal ym∈CN\mathbf{y}_m \in \mathbb{C}^N:

2bmNfs≀Cfhβ€…β€ŠβŸΉβ€…β€Šbm≀Cfh2Nfs2 b_m N f_s \leq C_{\text{fh}} \implies b_m \leq \frac{C_{\text{fh}}}{2 N f_s}

Since typically N>∣Dm∣N > |\mathcal{D}_m| for multi-antenna APs, Level 4 requires higher fronthaul capacity β€” or fewer quantization bits β€” than Levels 1–3.

Theorem: SINR Degradation Under Fronthaul Quantization

Under bb-bit uniform scalar quantization of the local estimates, the effective SINR of user kk with LSFD (Level 3) is

SINRk(3,q)=pk∣αkHbk∣2Ξ±kH(Ek+Οƒq2 diag(Ξ±k∘αkβˆ—))Ξ±k\text{SINR}_k^{(3,q)} = \frac{p_k |\boldsymbol{\alpha}_k^H \mathbf{b}_k|^2}{\boldsymbol{\alpha}_k^H \left( \mathbf{E}_k + \sigma^2_{q} \, \text{diag}(\boldsymbol{\alpha}_k \circ \boldsymbol{\alpha}_k^*) \right) \boldsymbol{\alpha}_k}

where Οƒq2\sigma^2_{q} is the per-component quantization noise variance. In the high-resolution regime (bβ‰₯5b \geq 5), the SINR loss relative to unquantized LSFD is approximately

Ξ”SINRkβ‰ˆΟƒq2βˆ₯Ξ±kβˆ₯2Ξ±kHEkΞ±kβ‰ˆc22b\Delta \text{SINR}_k \approx \frac{\sigma^2_{q} \|\boldsymbol{\alpha}_k\|^2}{\boldsymbol{\alpha}_k^H \mathbf{E}_k \boldsymbol{\alpha}_k} \approx \frac{c}{2^{2b}}

for a constant cc that depends on the signal dynamic range.

Quantization adds independent noise at each AP's fronthaul output. This noise is equivalent to increasing the thermal noise variance by Οƒq2\sigma^2_{q} per AP. With b=5b = 5 bits, 22b=10242^{2b} = 1024, so the quantization noise is roughly 30 dB below the signal β€” negligible in most scenarios. With b=3b = 3 bits, the noise is only 18 dB below, which can degrade the SINR by 1–2 dB.

Fronthaul Rate vs. Quantization Distortion

Explore the tradeoff between fronthaul rate (determined by quantization bits bb) and the resulting SINR degradation. Observe that 4–6 bits per dimension are sufficient for near-lossless performance in most scenarios.

Parameters
100
4
20
10

Example: Required Quantization Bits for 1 dB SINR Loss

A cell-free network operates at an average unquantized SINR of 15 dB. The dynamic range of the local estimates is s^maxβ‘βˆ’s^min⁑=4Οƒs\hat{s}_{\max} - \hat{s}_{\min} = 4\sigma_s where Οƒs2=0.1\sigma_s^2 = 0.1 is the variance of the local estimates. How many quantization bits bb are needed to limit the SINR degradation to at most 1 dB?

Definition:

Vector Quantization for Fronthaul Compression

Instead of quantizing each local estimate independently (scalar quantization), vector quantization jointly quantizes the vector s^m=[s^m1,…,s^m∣Dm∣]T\hat{\mathbf{s}}_m = [\hat{s}_{m1}, \ldots, \hat{s}_{m|\mathcal{D}_m|}]^T of all local estimates at AP mm. By exploiting the correlation structure (nearby users have correlated channels), vector quantization achieves the same distortion as scalar quantization with fewer bits.

The rate-distortion bound for Gaussian sources gives the minimum fronthaul rate:

Rfh,mβ‰₯βˆ‘i=1∣Dm∣max⁑(0,12log⁑2Ξ»iD)[bit/sample]R_{\text{fh},m} \geq \sum_{i=1}^{|\mathcal{D}_m|} \max\left(0, \frac{1}{2} \log_2 \frac{\lambda_i}{D}\right) \quad [\text{bit/sample}]

where Ξ»1β‰₯β‹―β‰₯λ∣Dm∣\lambda_1 \geq \cdots \geq \lambda_{|\mathcal{D}_m|} are the eigenvalues of the local estimate covariance matrix and DD is the target distortion level.

Vector quantization provides diminishing returns when the local estimates are nearly uncorrelated (well-separated users). The largest gain occurs when ∣Dm∣|\mathcal{D}_m| is large and users share similar channel subspaces.

Scalar vs. Vector Quantization for Fronthaul

AspectScalar QuantizationVector Quantization
Rate (bit/sample)2b∣Dm∣2b |\mathcal{D}_m|βˆ‘i12log⁑2(Ξ»i/D)\sum_i \frac{1}{2} \log_2(\lambda_i / D) (rate-distortion bound)
DistortionΞ”2/12\Delta^2 / 12 per componentOptimally distributed across eigenvalues (water-filling)
ComplexityO(∣Dm∣)O(|\mathcal{D}_m|)O(∣Dm∣2)O(|\mathcal{D}_m|^2) (covariance estimation + KLT)
Rate savingBaseline20–40% for correlated users
ImplementationSimple: standard ADCRequires eigendecomposition of local estimate covariance

Optimal Bit Allocation Across APs

Complexity: O(Mbmax⁑)O(M b_{\max}) where bmax⁑b_{\max} is the maximum allowed bits
Input: Fronthaul budget CfhC_{\text{fh}} per AP, user set Dm\mathcal{D}_m,
large-scale coefficients {Ξ²mk}\{\beta_{mk}\}, LSFD weights {Ξ±mk}\{\alpha_{mk}\}
Goal: Allocate quantization bits {bm}m=1M\{b_m\}_{m=1}^M to maximize the sum SINR
subject to per-AP fronthaul constraint.
1. Initialize: bm=⌊Cfh/(2∣Dm∣fs)βŒ‹b_m = \lfloor C_{\text{fh}} / (2 |\mathcal{D}_m| f_s) \rfloor for all mm
2. Compute the marginal SINR gain from adding one bit at each AP:
Ξ”m=SINRk(bm+1)βˆ’SINRk(bm)\Delta_m = \text{SINR}_k(b_m + 1) - \text{SINR}_k(b_m)
3. Greedy allocation: While fronthaul budget allows:
a. Find m⋆=arg⁑max⁑mΞ”mm^{\star} = \arg\max_m \Delta_m (AP with largest marginal gain)
b. If 2(bm⋆+1)∣Dmβ‹†βˆ£fs≀Cfh2(b_{m^{\star}} + 1) |\mathcal{D}_{m^{\star}}| f_s \leq C_{\text{fh}}:
set bm⋆←bm⋆+1b_{m^{\star}} \leftarrow b_{m^{\star}} + 1, update Ξ”m⋆\Delta_{m^{\star}}
c. Else: remove m⋆m^{\star} from candidates
4. Return {bm}\{b_m\}

The greedy algorithm gives a near-optimal solution because the SINR gain from each additional bit is a concave function of bmb_m (diminishing returns). APs with strong LSFD weights ∣αmk∣|\alpha_{mk}| receive more bits because their quantization noise has a larger impact on the final estimate.

🚨Critical Engineering Note

Fronthaul Technologies and Their Capacities

The choice of fronthaul technology determines the achievable cooperation level:

  • Dark fiber: 10–100 Gbps per link. Supports Level 4 with high-resolution quantization. Cost: high (fiber deployment), power: 0.5–2 W per link.
  • Ethernet (eCPRI): 10–25 Gbps per link. Supports Level 4 with moderate quantization or Level 3 with high resolution. The O-RAN standard mandates eCPRI for the fronthaul interface.
  • Millimeter-wave wireless: 1–10 Gbps per link. Supports Level 2–3. Used in urban small cells where fiber is not available.
  • Sub-6 GHz wireless: 100 Mbps–1 Gbps. Supports only Level 1–2 with aggressive compression.

In practice, heterogeneous fronthaul is common: some APs have fiber (enabling Level 4) while others use wireless backhaul (limited to Level 2–3). The cooperation level can be selected per-AP based on the available fronthaul capacity.

Practical Constraints
  • β€’

    eCPRI fronthaul adds 1–2 microseconds latency per hop

  • β€’

    Wireless fronthaul adds 10–50 microseconds, limiting real-time cooperation

  • β€’

    Fronthaul power scales with the data rate: approximately 0.1 W per Gbps

  • β€’

    Multi-hop fronthaul (daisy-chain topology) accumulates latency and limits cooperation

Common Mistake: Using the Additive Noise Model for Low-Resolution Quantization

Mistake:

Applying the additive quantization noise model (s^(q)=s^+e\hat{s}^{(q)} = \hat{s} + e with ee independent of s^\hat{s}) when b≀3b \leq 3 bits.

Correction:

The additive noise model is accurate only for bβ‰₯4b \geq 4–55 bits. For very low resolution (b=1b = 1–33), the quantization error is correlated with the input and the Bussgang decomposition should be used instead: s^(q)=Ξ±Bs^+eB\hat{s}^{(q)} = \alpha_{\text{B}} \hat{s} + e_{\text{B}} where Ξ±B=1βˆ’2βˆ’2b/3\alpha_{\text{B}} = 1 - 2^{-2b}/3 is the Bussgang gain and eBe_{\text{B}} is uncorrelated with s^\hat{s}. The Bussgang model accounts for the signal attenuation caused by coarse quantization.

Quick Check

An AP with N=4N = 4 antennas has a 10 Gbps eCPRI fronthaul link. The system bandwidth is fs=100f_s = 100 MHz. What is the maximum number of quantization bits per real dimension for Level 4 operation?

b=6b = 6 bits

b=2b = 2 bits

b=16b = 16 bits

b=1b = 1 bit

The Fronthaul Perspective on Cooperation Levels

The fronthaul analysis provides a unified view of the cooperation levels:

  • Level 1–2 (local combining): Fronthaul carries ∣Dm∣|\mathcal{D}_m| complex scalars per channel use. With 4–6 bit quantization and ∣Dm∣=10|\mathcal{D}_m| = 10: 0.8–1.2 Gbps per AP.
  • Level 3 (LSFD): Same fronthaul as Level 2, plus LSFD weights updated every 100–1000 ms (negligible additional overhead).
  • Level 4 (centralized MMSE): Fronthaul carries NN complex vectors per channel use. With 6–8 bit quantization and N=4N = 4: 4.8–6.4 Gbps per AP.

This quantifies the fundamental tradeoff: Level 3 provides 70–90% of Level 4's SINR performance at 15–25% of the fronthaul cost. When fronthaul capacity is the bottleneck β€” which it is in most deployments β€” Level 3 is the rational choice.

Key Takeaway

Fronthaul quantization with 4–6 bits per dimension is sufficient for near-lossless distributed processing. The SINR loss from quantization is less than 0.5 dB in most scenarios, which is far smaller than the gap between cooperation levels. The system designer should choose the cooperation level first (based on fronthaul capacity and AP antennas), then allocate quantization bits within the chosen level. Vector quantization provides 20–40% rate savings over scalar quantization when users are spatially correlated, but the added complexity is justified only in fronthaul-constrained deployments.

Fronthaul Compression

The process of quantizing and encoding the signals transmitted between access points and the central processing unit over capacity-limited fronthaul links. Includes scalar quantization, vector quantization, and information-theoretic compression based on Wyner-Ziv coding.

Related: Fronthaul, Scalar Quantization of Local Estimates, Cooperation Level

Bussgang Decomposition

A technique for analyzing nonlinear systems (such as quantizers) applied to Gaussian inputs. The output is decomposed as Q(x)=Ξ±x+eQ(x) = \alpha x + e where Ξ±\alpha is a deterministic gain and ee is uncorrelated with xx. Used to analyze low-resolution ADCs and fronthaul quantization in massive MIMO systems.

Related: Scalar Quantization of Local Estimates, Fronthaul Compression