Manifold Optimization on the Complex Torus
Respect the Geometry
SDR lifts the problem to a convex space; element-wise BCD cycles coordinates. Manifold optimization takes a different view: it respects the native geometry of the feasible set. The unit-modulus torus is a smooth manifold , and standard tools from Riemannian geometry (gradients, retractions, geodesics) give convergent, efficient algorithms tailored to this space. Manifold methods scale much better than SDR and give better solutions than pure element-wise. For large (), manifold is the algorithm of choice.
Definition: The Complex Unit-Modulus Manifold
The Complex Unit-Modulus Manifold
The complex unit-modulus manifold is
is the -fold Cartesian product of complex unit circles: . Each circle is a 1D smooth manifold, so is an -dimensional smooth manifold.
The tangent space at a point is
Each tangent vector is "perpendicular" to in a pointwise sense β it advances along the circle at each coordinate without leaving the torus.
The condition is the linearization of : differentiating gives . The tangent space is thus the orthogonal complement of in the per-coordinate sense.
Definition: Riemannian Gradient and Retraction
Riemannian Gradient and Retraction
Given a smooth and its Euclidean gradient , the Riemannian gradient at is the projection of onto the tangent space:
where is element-wise multiplication.
A retraction maps a tangent vector back to a point on . A simple choice is per-coordinate normalization:
More sophisticated retractions use the exponential map (exact geodesic), but the simple normalization is typically sufficient.
Riemannian Gradient Descent on
Complexity: : - iterations, each for gradient of quadraticThe per-iteration cost is dominated by the gradient of the quadratic, which is for dense . Compared to SDR's , manifold is times faster β a decisive advantage at . Conjugate gradient or L-BFGS variants on the manifold are used in practice for faster convergence (see Manopt toolbox).
Theorem: Convergence of Riemannian Gradient Descent
Assume is twice continuously differentiable on , and the manifold is compact (trivially, as the complex torus). Riemannian gradient ascent with Armijo backtracking satisfies:
- unless .
- The sequence has a convergent subsequence whose limit is a stationary point of on (i.e., ).
- Near a non-degenerate stationary point, convergence is linear with rate governed by the Riemannian Hessian.
For the RIS QCQP, the method typically converges to a local maximum within 50β200 iterations.
The Riemannian gradient points in the direction of steepest ascent on the manifold. Armijo backtracking ensures monotone improvement. Under smoothness, the sequence converges to a stationary point β the Riemannian analog of a zero gradient.
Riemannian Newton and Trust Region
Riemannian gradient descent converges linearly; Riemannian Newton converges quadratically β each iteration squares the optimality gap. Computing the Riemannian Hessian (the second fundamental form of + Euclidean Hessian) is , same as the gradient, so Newton trades no compute per iteration for faster outer convergence. Practical implementations use the trust-region Newton method via the Manopt toolbox, giving the fastest reliable convergence for the RIS QCQP.
Riemannian Gradient Descent on
Example: Riemannian Gradient for the Rank-1 QCQP
For , compute the Euclidean and Riemannian gradients at (all ones).
Euclidean gradient
at .
Riemannian gradient
Project to tangent space: . At : , so .
Interpretation
At , the Riemannian gradient is purely imaginary β a tangent vector in the rotational direction at each coordinate. This makes intuitive sense: the tangent space at is the imaginary axis at each coordinate, since advancing on the unit circle at means moving purely in the imaginary direction.
Manifold GD vs. SDR Convergence
Track the objective of Riemannian gradient descent (with Armijo backtracking) over iterations for a fixed problem instance, and compare with the SDR upper bound and the element-wise BCD trajectory. Manifold typically closes 80-90% of the gap to SDR within 50 iterations; element-wise plateaus earlier but at a lower objective.
Parameters
Common Mistake: Step Size Matters
Mistake:
"Just use a fixed step size and iterate."
Correction:
Without backtracking, a fixed step size can cause non-monotone iterates and even divergence (overshooting). Always use Armijo backtracking: start with , halve it until for some . This guarantees monotone progress at the cost of a few extra function evaluations per iteration. Modern implementations (Manopt) handle this automatically.
Manopt: Let Someone Else Do the Work
The Manopt toolbox (MATLAB) and its Python/Julia ports implement Riemannian trust-region Newton out of the box for the complex unit-modulus manifold (among many others). Using Manopt:
- Define the objective and its Euclidean gradient.
- Select the manifold ("complexcircle" or "obliquecomplexfactor").
- Call
trustregions(problem).
The toolbox handles tangent-space projection, retraction, Hessian approximation, and backtracking internally. For research and production alike, Manopt is the recommended black-box; avoid re-implementing Riemannian optimization unless you have a very specific reason.
- β’
Manopt at : typically 5-10 s per solve, iterations.
- β’
Pymanopt / Julia Manopt have identical API β easy to port across languages.
- β’
Warm-starting across AO iterations reduces per-solve time by -.