Sum-Rate Maximization

The Hardest Fairness Criterion

Sum-rate maximization β€” choosing powers to maximize βˆ‘kRk\sum_k R_k β€” is the most natural objective from a throughput perspective but also the most challenging to optimize. Unlike max-min fairness (quasi-concave) and proportional fairness (GP-amenable), the sum-rate objective is non-concave in the power vector due to the interference coupling. This section develops two powerful algorithmic frameworks: successive convex approximation (SCA) and the weighted minimum mean-square error (WMMSE) algorithm.

Definition:

Weighted Sum-Rate Maximization

The weighted sum-rate maximization (WSRMax) problem is

max⁑pβ‰₯0β€…β€Šβˆ‘k=1Kwklog⁑2 ⁣(1+SINRk(p))s.t.βˆ‘k=1Kpk≀Pt\max_{\mathbf{p} \geq \mathbf{0}} \; \sum_{k=1}^{K} w_k \log_2\!\left(1 + \text{SINR}_k(\mathbf{p})\right) \quad \text{s.t.} \quad \sum_{k=1}^{K} p_k \leq P_t

where wk>0w_k > 0 are user-specific weights. When wk=1w_k = 1 for all kk, this is the unweighted sum-rate problem. By varying the weights, one can trace the boundary of the achievable rate region.

The weights wkw_k can encode priority (premium users get higher weight), queue state (users with more buffered data get higher weight), or fairness constraints (setting wk∝1/RΛ‰kw_k \propto 1/\bar{R}_k approximates proportional fairness).

Theorem: NP-Hardness of Sum-Rate Maximization

The weighted sum-rate maximization problem with interference is NP-hard in general. Specifically, even for the two-user interference channel with single-antenna nodes, the problem of finding the power allocation maximizing w1R1+w2R2w_1 R_1 + w_2 R_2 is NP-hard when both users have non-zero interference coupling.

The difficulty arises from the non-concavity of the sum-rate function. Increasing pkp_k improves RkR_k (more signal) but hurts RlR_l for l≠kl \neq k (more interference). This creates a landscape with potentially many local maxima, and no polynomial-time algorithm is known to find the global optimum.

Successive Convex Approximation (SCA)

An iterative optimization technique that replaces a non-convex problem with a sequence of convex subproblems. Each subproblem is constructed by approximating the non-convex parts with concave lower bounds, and the iterates provably converge to a stationary point of the original problem.

Related: WMMSE Algorithm

Definition:

Successive Convex Approximation for Sum-Rate

The SCA approach to WSRMax works by lower-bounding the concave part of RkR_k and linearizing the interference term. At iteration nn, given the current powers p(n)\mathbf{p}^{(n)}, we solve

max⁑pβ‰₯0β€…β€Šβˆ‘k=1KwkR~k(n)(p)s.t.βˆ‘kpk≀Pt\max_{\mathbf{p} \geq \mathbf{0}} \; \sum_{k=1}^{K} w_k \tilde{R}_k^{(n)}(\mathbf{p}) \quad \text{s.t.} \quad \sum_k p_k \leq P_t

where R~k(n)\tilde{R}_k^{(n)} is a concave surrogate of RkR_k satisfying: (i) R~k(n)(p(n))=Rk(p(n))\tilde{R}_k^{(n)}(\mathbf{p}^{(n)}) = R_k(\mathbf{p}^{(n)}) (tight at current point), (ii) βˆ‡R~k(n)(p(n))=βˆ‡Rk(p(n))\nabla \tilde{R}_k^{(n)}(\mathbf{p}^{(n)}) = \nabla R_k(\mathbf{p}^{(n)}) (matching gradient), and (iii) R~k(n)(p)≀Rk(p)\tilde{R}_k^{(n)}(\mathbf{p}) \leq R_k(\mathbf{p}) for all p\mathbf{p} (lower bound). Properties (i)–(iii) guarantee that the sequence {p(n)}\{\mathbf{p}^{(n)}\} converges to a stationary point.

Theorem: Convergence of SCA

Let {p(n)}nβ‰₯0\{\mathbf{p}^{(n)}\}_{n \geq 0} be the sequence of iterates produced by the SCA algorithm. If the surrogate functions R~k(n)\tilde{R}_k^{(n)} satisfy properties (i)–(iii) above, then:

  1. The objective is non-decreasing: βˆ‘kwkRk(p(n+1))β‰₯βˆ‘kwkRk(p(n))\sum_k w_k R_k(\mathbf{p}^{(n+1)}) \geq \sum_k w_k R_k(\mathbf{p}^{(n)}).
  2. Every limit point of {p(n)}\{\mathbf{p}^{(n)}\} satisfies the KKT conditions of the original WSRMax problem.

The key idea is that each SCA iteration maximizes a lower bound that is tight at the current point. Because the bound is tight, the new iterate achieves at least the same objective value. Because the gradient matches, the stationary points of the surrogates coincide with those of the original problem.

,

Definition:

The WMMSE Algorithm

The Weighted Minimum Mean-Square Error (WMMSE) algorithm is a particular instance of SCA that exploits the SINR–MSE relationship. For user kk with receive filter gkg_k and weight ΞΌk\mu_k, define the MSE

ek=E[∣gkykβˆ’sk∣2]=∣gk∣2(βˆ‘l=1Kpl∣vkHHl∣2+Οƒ2)βˆ’2Re(gkpkvkHHk)+1.e_k = \mathbb{E}\left[|g_k y_k - s_k|^2\right] = |g_k|^2 \left(\sum_{l=1}^{K} p_l |\mathbf{v}_k^H \mathbf{H}_{l}|^2 + \sigma^2\right) - 2\text{Re}(g_k \sqrt{p_k} \mathbf{v}_k^H \mathbf{H}_{k}) + 1.

The WMMSE algorithm alternates between:

  1. Fix p\mathbf{p}, update gkg_k: MMSE receiver gk⋆=pkvkHHkβˆ‘lpl∣vkHHl∣2+Οƒ2g_k^\star = \frac{\sqrt{p_k}\mathbf{v}_k^H\mathbf{H}_{k}}{\sum_l p_l |\mathbf{v}_k^H\mathbf{H}_{l}|^2 + \sigma^2}
  2. Fix gkg_k, update ΞΌk\mu_k: Weight ΞΌk⋆=1/ek⋆\mu_k^\star = 1/e_k^\star (inverse MSE)
  3. Fix gk,ΞΌkg_k, \mu_k, update p\mathbf{p}: Solve a convex quadratic program

This alternating optimization provably converges to a stationary point of the WSRMax problem.

WMMSE Algorithm

The Weighted Minimum Mean-Square Error algorithm for sum-rate maximization. It alternates between updating MMSE receivers, MSE weights, and transmit powers, converging to a KKT point of the weighted sum-rate problem.

Related: Successive Convex Approximation for Sum-Rate, Sum Rate Maximization

WMMSE Algorithm for Weighted Sum-Rate Maximization

Complexity: O(Iβ‹…K2)O(I \cdot K^{2}) per iteration; typically I=10I = 10–5050 iterations to converge
Input: Channel vectors {Hk}\{\mathbf{H}_{k}\}, combining vectors {vk}\{\mathbf{v}_k\},
weights {wk}\{w_k\}, power budget PtP_t, tolerance Ο΅\epsilon
Output: Power vector p⋆\mathbf{p}^\star (stationary point of WSRMax)
1. Initialize p(0)\mathbf{p}^{(0)} (e.g., equal power)
2. repeat
3. \quad for k=1,…,Kk = 1, \ldots, K do
4. gk←pkvkHHkβˆ‘lpl∣vkHHl∣2+Οƒ2\quad\quad g_k \leftarrow \frac{\sqrt{p_k} \mathbf{v}_k^H \mathbf{H}_{k}}{\sum_l p_l |\mathbf{v}_k^H \mathbf{H}_{l}|^2 + \sigma^2} (MMSE receiver)
5. ek←1βˆ’βˆ£gk∣2pk∣vkHHk∣2/(βˆ‘lpl∣vkHHl∣2+Οƒ2)\quad\quad e_k \leftarrow 1 - |g_k|^2 p_k |\mathbf{v}_k^H \mathbf{H}_{k}|^2 / (\sum_l p_l |\mathbf{v}_k^H \mathbf{H}_{l}|^2 + \sigma^2)
6. ΞΌk←wk/ek\quad\quad \mu_k \leftarrow w_k / e_k (MSE weight)
7. \quad end for
8. \quad Solve for p\mathbf{p}: min⁑pβ‰₯0, 1Tp≀Ptβˆ‘kΞΌkek(p)βˆ’wklog⁑(ΞΌkek(p))\min_{\mathbf{p} \geq 0, \, \mathbf{1}^T\mathbf{p} \leq P_t} \sum_k \mu_k e_k(\mathbf{p}) - w_k \log(\mu_k e_k(\mathbf{p}))
9. until βˆ£Ξ”WSR∣<Ο΅|\Delta \text{WSR}| < \epsilon

The key insight of WMMSE is the identity max⁑μk>0(wklog⁑μkβˆ’wkΞΌkek+wk)=wklog⁑(1/ekmin⁑)\max_{\mu_k > 0}(w_k \log \mu_k - w_k \mu_k e_k + w_k) = w_k \log(1/e_k^{\min}) which connects the MSE to the rate. This allows the non-convex rate maximization to be decomposed into alternating convex steps.

Example: WMMSE Convergence for a 4-User System

Consider Nt=64N_t = 64 antennas, K=4K = 4 users with random i.i.d. Rayleigh channels, ZF combining, and SNR=10\text{SNR} = 10 dB. Run the WMMSE algorithm from equal-power initialization and compare the converged sum rate to the equal-power baseline.

Common Mistake: WMMSE Finds Local Optima, Not Global

Mistake:

A common error is to trust that the WMMSE algorithm finds the global optimum of the sum-rate problem because it is based on alternating optimization with closed-form updates.

Correction:

WMMSE converges to a stationary point (KKT point) of the sum-rate problem, which is a local optimum or saddle point β€” not necessarily the global optimum. In practice, running WMMSE from multiple random initializations and taking the best result improves the chance of finding a good solution. For massive MIMO with favorable propagation, the interference coupling is weak and the landscape has few local optima, so a single run usually suffices.

WMMSE Algorithm Convergence

Observe how the weighted sum rate evolves over WMMSE iterations. Compare convergence speed for different numbers of users and antenna counts.

Parameters
4
64
10
30
⚠️Engineering Note

Computational Cost of WMMSE in Real-Time Systems

The WMMSE algorithm requires 10–50 iterations, each involving matrix inversions of size KΓ—KK \times K and a convex power allocation step. For massive MIMO with K=16K = 16 users and 1 ms scheduling granularity (5G NR), the algorithm must complete within approximately 100 microseconds on the baseband processor. This is achievable for small KK but becomes challenging for dense deployments with 64+ users.

Practical Constraints
  • β€’

    5G NR scheduling period: 1 ms (2 slots at 30 kHz SCS)

  • β€’

    Baseband processing budget: typically 100-200 microseconds for power allocation

  • β€’

    WMMSE with K=16K = 16 and 30 iterations: ~50 microseconds on a modern DSP

  • β€’

    For K>32K > 32, heuristic methods from Section 5.4 are preferred

Quick Check

What is the key mathematical identity that enables the WMMSE algorithm?

The rate Rk=log⁑(1+SINRk)R_k = \log(1 + \text{SINR}_k) equals βˆ’log⁑(ekmin⁑)-\log(e_k^{\min}) where ekmin⁑e_k^{\min} is the MMSE

The sum rate is concave in the power vector under favorable propagation

The interference term can be linearized without loss of optimality

Water-filling is optimal when treating interference as noise

Key Takeaway

Weighted sum-rate maximization is NP-hard in general but yields to iterative methods. The WMMSE algorithm β€” which alternates between MMSE receiver updates, weight updates, and convex power allocation β€” converges to a KKT point and is the standard workhorse for sum-rate optimization in massive MIMO systems.