Chapter Summary

Chapter 13 Summary

Key Points

  • 1.

    The sparse recovery problem is underdetermined but not hopeless. With y=Ax+w∈RM\mathbf{y} = \mathbf{A}\mathbf{x}+\mathbf{w}\in\mathbb{R}^M, A∈RMΓ—N\mathbf{A}\in\mathbb{R}^{M\times N}, Mβ‰ͺNM \ll N, the linear system has infinitely many solutions. A sparsity prior (βˆ₯xβˆ₯0≀s\|\mathbf{x}\|_0 \leq s) restores identifiability: when s<spark(A)/2s < \mathrm{spark}(\mathbf{A})/2, the ss-sparse solution is unique.

  • 2.

    β„“0\ell_0 is NP-hard; β„“1\ell_1 is its convex hull. Natarajan (1995) proved the computational hardness of β„“0\ell_0-minimization. The β„“1\ell_1 relaxation replaces the combinatorial penalty by its tightest convex surrogate. Geometrically, the β„“1\ell_1 unit ball is a cross-polytope whose vertices coincide with 11-sparse signals β€” minimizing βˆ₯xβˆ₯1\|\mathbf{x}\|_1 on an affine subspace is most likely to land on a vertex, hence on a sparse solution.

  • 3.

    Basis Pursuit and LASSO are equivalent for suitable parameters. BPDN minimizes βˆ₯xβˆ₯1\|\mathbf{x}\|_1 s.t.\ βˆ₯yβˆ’Axβˆ₯2≀ϡ\|\mathbf{y}-\mathbf{A}\mathbf{x}\|_2\leq\epsilon; LASSO minimizes 12βˆ₯yβˆ’Axβˆ₯22+Ξ»βˆ₯xβˆ₯1\tfrac{1}{2}\|\mathbf{y}-\mathbf{A}\mathbf{x}\|_2^2+\lambda\|\mathbf{x}\|_1. They trace the same solution path as Ο΅\epsilon and Ξ»\lambda vary. Both are convex, hence solvable globally and efficiently by modern algorithms (Ch 14).

  • 4.

    The RIP is the right stability condition. A\mathbf{A} satisfies RIP of order ss with constant Ξ΄s\delta_s when (1βˆ’Ξ΄s)βˆ₯xβˆ₯22≀βˆ₯Axβˆ₯22≀(1+Ξ΄s)βˆ₯xβˆ₯22(1-\delta_s)\|\mathbf{x}\|_2^2 \leq \|\mathbf{A}\mathbf{x}\|_2^2 \leq (1+\delta_s)\|\mathbf{x}\|_2^2 for every ss-sparse x\mathbf{x}. If Ξ΄2s<2βˆ’1\delta_{2s} < \sqrt{2}-1, BPDN recovers every ss-sparse x\mathbf{x} stably. Verifying RIP is NP-hard, but random Gaussian/Bernoulli matrices satisfy it with overwhelming probability when M=O(slog⁑(N/s))M = O(s\log(N/s)).

  • 5.

    Coherence is a cheap surrogate for RIP. The mutual coherence ΞΌ(A)=max⁑iβ‰ j∣⟨Ai,Aj⟩∣/(βˆ₯Aiβˆ₯βˆ₯Ajβˆ₯)\mu(\mathbf{A})=\max_{i\neq j}|\langle\mathbf{A}_{i},\mathbf{A}_{j}\rangle|/(\|\mathbf{A}_{i}\|\|\mathbf{A}_{j}\|) can be computed in O(MN2)O(MN^2). The Welch bound gives ΞΌβ‰₯(Nβˆ’M)/(M(Nβˆ’1))\mu \geq \sqrt{(N-M)/(M(N-1))}. Recovery holds when s<12(1+1/ΞΌ)s < \tfrac{1}{2}(1+1/\mu), a weaker but easier-to-check guarantee than RIP.

  • 6.

    Recovery bounds degrade gracefully under noise. Noiseless RIP gives exact recovery; with noise βˆ₯wβˆ₯2≀ϡ\|\mathbf{w}\|_2\leq\epsilon, BPDN returns x^\hat{\mathbf{x}} with βˆ₯x^βˆ’x⋆βˆ₯2≀CΟ΅+Cβ€²Οƒs(x⋆)1/s\|\hat{\mathbf{x}}-\mathbf{x}^\star\|_2 \leq C\epsilon + C'\sigma_s(\mathbf{x}^\star)_1/\sqrt{s}. The first term is unavoidable noise propagation; the second penalises non-exact sparsity (Οƒs(x⋆)1=\sigma_s(\mathbf{x}^\star)_1 = best ss-term approximation error). Oracle inequalities for LASSO show βˆ₯x^βˆ’x⋆βˆ₯22=O(slog⁑Nβ‹…Οƒ2)\|\hat{\mathbf{x}}-\mathbf{x}^\star\|_2^2 = O(s\log N\cdot \sigma^2).

Looking Ahead

Chapter 14 develops the algorithms that actually solve BPDN / LASSO at scale: ISTA, FISTA, ADMM, OMP, and Bayesian alternatives. Chapter 15 applies this machinery to wireless problems β€” sparse channel estimation, massive activity detection, and DOA estimation β€” where compressed sensing has become the workhorse of CommIT research.