Convex Optimization Problems

Definition:

Standard-Form Optimization Problem

A general optimization problem in standard form is

minimisef0(x)subject tofi(x)0,i=1,,mhj(x)=0,j=1,,p\begin{aligned} \text{minimise} \quad & f_0(\mathbf{x}) \\ \text{subject to} \quad & f_i(\mathbf{x}) \leq 0, \quad i = 1, \ldots, m \\ & h_j(\mathbf{x}) = 0, \quad j = 1, \ldots, p \end{aligned}

where xRn\mathbf{x} \in \mathbb{R}^n is the optimisation variable, f0f_0 is the objective, fif_i are inequality constraints, and hjh_j are equality constraints.

The optimal value is p=inf{f0(x):x feasible}p^\star = \inf\{f_0(\mathbf{x}) : \mathbf{x} \text{ feasible}\}.

If the problem is a maximisation, we convert it by minimising f0-f_0.

Definition:

Convex Optimization Problem

An optimization problem is convex if:

  1. The objective f0f_0 is convex.
  2. Each inequality constraint function fif_i is convex.
  3. Each equality constraint is affine: hj(x)=ajTxbjh_j(\mathbf{x}) = \mathbf{a}_j^T \mathbf{x} - b_j.

Equivalently, the feasible set is convex and the objective is convex over it.

The fundamental property: every local minimum of a convex problem is a global minimum.

Theorem: Local Optima Are Global Optima

For a convex optimization problem, any locally optimal point is also globally optimal.

If there were a better point elsewhere, the line segment connecting it to the local minimum would contain a feasible point with even lower objective, contradicting local optimality.

Definition:

Linear Program (LP)

A linear program has a linear objective and linear constraints:

minimisecTxsubject toAxb,Cx=d\begin{aligned} \text{minimise} \quad & \mathbf{c}^T \mathbf{x} \\ \text{subject to} \quad & \mathbf{A}\mathbf{x} \preceq \mathbf{b}, \quad \mathbf{C}\mathbf{x} = \mathbf{d} \end{aligned}

LPs are convex and solvable in polynomial time (e.g., interior-point methods). The feasible set is a polyhedron, and the optimum is attained at a vertex (extreme point) when bounded.

Definition:

Quadratic Program (QP)

A quadratic program has a quadratic objective and linear constraints:

minimise12xTQx+cTxsubject toAxb\begin{aligned} \text{minimise} \quad & \tfrac{1}{2}\mathbf{x}^T \mathbf{Q}\mathbf{x} + \mathbf{c}^T \mathbf{x} \\ \text{subject to} \quad & \mathbf{A}\mathbf{x} \preceq \mathbf{b} \end{aligned}

The QP is convex if and only if Q0\mathbf{Q} \succeq 0.

Definition:

Second-Order Cone Program (SOCP)

A second-order cone program has the form:

minimisefTxsubject toAix+bi2ciTx+di,i=1,,m\begin{aligned} \text{minimise} \quad & \mathbf{f}^T \mathbf{x} \\ \text{subject to} \quad & \|\mathbf{A}_i \mathbf{x} + \mathbf{b}_i\|_2 \leq \mathbf{c}_i^T \mathbf{x} + d_i, \quad i = 1, \ldots, m \end{aligned}

SOCPs generalise LPs and QPs and are solvable in polynomial time.

Robust beamforming problems in MIMO systems are naturally SOCPs: the constraint wPmax1/2\|\mathbf{w}\| \leq P_{\max}^{1/2} is a second-order cone constraint.

Definition:

Semidefinite Program (SDP)

A semidefinite program optimises a linear objective over the intersection of the PSD cone with affine constraints:

minimisetr(CX)subject totr(AiX)=bi,i=1,,mX0\begin{aligned} \text{minimise} \quad & \text{tr}(\mathbf{C}\mathbf{X}) \\ \text{subject to} \quad & \text{tr}(\mathbf{A}_i \mathbf{X}) = b_i, \quad i = 1, \ldots, m \\ & \mathbf{X} \succeq 0 \end{aligned}

SDPs generalise LPs, QPs, and SOCPs. They are solvable in polynomial time via interior-point methods.

SDPs appear throughout wireless communications: relaxations of beamformer design, MIMO transceiver optimisation, and sensor localisation.

Hierarchy of Convex Programs

ProblemObjectiveConstraintsExample in Wireless
LPLinearLinearPower control (high-SNR regime)
QPQuadratic (Q0\mathbf{Q} \succeq 0)LinearMMSE receiver design
SOCPLinearSecond-order coneRobust beamforming
SDPLinear in X\mathbf{X}X0\mathbf{X} \succeq 0 + affineMIMO transmit covariance optimisation

Definition:

Lagrangian and Lagrangian Dual Function

For the standard-form problem, the Lagrangian is

L(x,λ,ν)=f0(x)+i=1mλifi(x)+j=1pνjhj(x)\mathcal{L}(\mathbf{x}, \boldsymbol{\lambda}, \boldsymbol{\nu}) = f_0(\mathbf{x}) + \sum_{i=1}^m \lambda_i f_i(\mathbf{x}) + \sum_{j=1}^p \nu_j h_j(\mathbf{x})

where λ0\boldsymbol{\lambda} \succeq 0 are the dual variables (or Lagrange multipliers) for the inequality constraints and ν\boldsymbol{\nu} are the multipliers for equalities.

The Lagrangian dual function is

g(λ,ν)=infxL(x,λ,ν).g(\boldsymbol{\lambda}, \boldsymbol{\nu}) = \inf_{\mathbf{x}} \mathcal{L}(\mathbf{x}, \boldsymbol{\lambda}, \boldsymbol{\nu}).

The dual function is always concave (as the pointwise infimum of affine functions of (λ,ν)(\boldsymbol{\lambda}, \boldsymbol{\nu})), regardless of convexity of the original problem.

Theorem: Weak and Strong Duality

Weak duality: For any λ0\boldsymbol{\lambda} \succeq 0 and any ν\boldsymbol{\nu}, g(λ,ν)pg(\boldsymbol{\lambda}, \boldsymbol{\nu}) \leq p^\star.

The dual problem is d=supλ0,νg(λ,ν)d^\star = \sup_{\boldsymbol{\lambda} \succeq 0, \boldsymbol{\nu}} g(\boldsymbol{\lambda}, \boldsymbol{\nu}), so dpd^\star \leq p^\star always.

Strong duality (d=pd^\star = p^\star) holds for convex problems that satisfy a constraint qualification, the most common being Slater's condition: there exists a strictly feasible point x~\tilde{\mathbf{x}} with fi(x~)<0f_i(\tilde{\mathbf{x}}) < 0 for all i=1,,mi = 1, \ldots, m.

Weak duality says the dual always provides a lower bound on the primal optimal value. Slater's condition ensures the bound is tight. In wireless problems, strong duality almost always holds because the constraints typically have non-empty interior.

Lagrangian Dual Function for a Simple QP

Visualise the Lagrangian dual function g(λ)g(\lambda) for the quadratic program min  x2  s.t.  xa\min\; x^2 \;\text{s.t.}\; x \geq a. The dual provides a lower bound on the optimal value, and the duality gap closes when strong duality holds.

Parameters
1

Example: Dual of a Linear Program

Find the dual of the LP: min  cTx\min\;\mathbf{c}^T \mathbf{x} subject to Axb\mathbf{A}\mathbf{x} \preceq \mathbf{b}.

Why This Matters: SDPs in MIMO Beamforming

The optimal downlink beamforming problem (minimise total transmit power subject to per-user SINR constraints) can be formulated as an SOCP or, via rank-one relaxation Wk=wkwkH\mathbf{W}_k = \mathbf{w}_k \mathbf{w}_k^H, as an SDP. The SDP relaxation is tight (the optimal Wk\mathbf{W}_k is always rank one), so the relaxation solves the original problem exactly. This remarkable result, due to Bengtsson and Ottersten (2001), is a cornerstone of modern beamforming theory.

See full treatment in MMSE-SIC Optimality

Hierarchy of Convex Programs

Hierarchy of Convex Programs
The nesting of convex program classes: LP \subset QP \subset SOCP \subset SDP \subset general convex. Each class includes all classes above it and adds expressiveness.
⚠️Engineering Note

Numerical Precision of Interior-Point Solvers

Modern interior-point solvers (MOSEK, SeDuMi, SDPT3) solve SDPs and SOCPs to roughly 6–8 digits of accuracy in O(n3.5)O(n^{3.5}) time. In practice, be aware of the following:

  • Conditioning: Poorly scaled problems (coefficients spanning many orders of magnitude) cause numerical issues. Always normalise constraints so that coefficients are O(1)O(1).
  • Solver tolerances: Default feasibility/optimality tolerances are typically 10810^{-8}. For beamforming problems with per-antenna power constraints, tighter tolerances (101010^{-10}) may be needed to ensure the rank-one property of SDP solutions.
  • Problem size limits: SDPs with matrix variable size n>500n > 500 become impractical for general-purpose solvers. First-order methods (ADMM, proximal) scale better but give lower accuracy.
  • Standard tools: CVXPY (Python) and CVX (MATLAB) provide disciplined convex programming interfaces that automatically verify convexity and select appropriate solvers.
Practical Constraints
  • SDP matrix variable size limited to ~500 for general-purpose interior-point solvers

  • Problem conditioning: normalise coefficients to O(1)O(1)

  • Default solver tolerance 108\sim 10^{-8}; tighten for rank-constrained problems

📋 Ref: MOSEK, SeDuMi, SDPT3 solver documentation

Common Mistake: Not All QPs Are Convex

Mistake:

Assuming that any quadratic program is convex and efficiently solvable.

Correction:

A QP is convex only if Q0\mathbf{Q} \succeq 0. If Q\mathbf{Q} has negative eigenvalues, the problem is non-convex and generally NP-hard. Always check the Hessian/Q matrix before claiming convexity.

Lagrangian

The function L(x,λ,ν)=f0(x)+iλifi(x)+jνjhj(x)\mathcal{L}(\mathbf{x}, \boldsymbol{\lambda}, \boldsymbol{\nu}) = f_0(\mathbf{x}) + \sum_i \lambda_i f_i(\mathbf{x}) + \sum_j \nu_j h_j(\mathbf{x}) that incorporates constraints into the objective via dual variables.

Related: Dual Function, KKT Conditions

Duality Gap

The difference pdp^\star - d^\star between the primal and dual optimal values. Under Slater's condition, the gap is zero (strong duality).

Related: Lagrangian and Lagrangian Dual Function, Slater Condition

Slater's Condition

A constraint qualification for convex problems: there exists a point x~\tilde{\mathbf{x}} that is strictly feasible, meaning fi(x~)<0f_i(\tilde{\mathbf{x}}) < 0 for all inequality constraints. Guarantees strong duality (p=dp^\star = d^\star).

Related: Duality Gap, Weak and Strong Duality