Exercises
ex-ch13-01
EasyShow that the "norm" is not a norm (it fails the positive-homogeneity axiom).
Test for .
Scale invariance
For , . A norm requires , which fails.
ex-ch13-02
EasyCompute and for . Compare with .
Apply the definitions directly.
Compute
, , .
ex-ch13-03
EasyWrite down the soft-thresholding operator and evaluate it at for .
.
Evaluations
.
ex-ch13-04
EasyA dictionary has unit-norm columns with for all . What is the largest for which coherence-based recovery guarantees sparse recovery?
Use .
Plug in
, so . Exact recovery is guaranteed for .
ex-ch13-05
EasyState the Welch lower bound on coherence for a unit-norm dictionary.
.
Compute
.
ex-ch13-06
MediumProve that the -ball in has exactly vertices, and identify them.
Vertices correspond to -sparse signals.
Characterize vertices
A vertex is an extreme point: a point that is not a convex combination of two others in the set. In the -ball, these are for .
Count
There are such points, all -sparse with .
ex-ch13-07
MediumLet where is the Hadamard matrix scaled to unit-norm columns. Compute the coherence.
Columns of vs columns of : inner product is .
Worst-case pair
Any column of paired with any column of has inner product . Columns of are pairwise orthogonal; columns of are pairwise orthogonal.
Coherence
, achieving the Welch bound for .
ex-ch13-08
MediumShow that LASSO is a convex optimization problem. Identify the role of convexity.
Both terms are convex. Sum of convex is convex.
Quadratic term
is a convex quadratic in (Hessian ).
$\ell_1$ norm
Every norm is convex by the triangle inequality.
Convexity matters
Every local minimum is global. Descent algorithms converge to the global optimum. The dual is well-posed and the KKT conditions characterize optima. Convexity is why LASSO scales to millions of variables.
ex-ch13-09
MediumLet satisfy RIP with . Show that every -sparse vector is uniquely determined from .
implies injectivity on -sparse vectors.
Suppose two distinct $s$-sparse recoveries
If with both -sparse, then is -sparse and .
Contradiction
RIP gives , contradicting .
ex-ch13-10
MediumDerive the KKT conditions for LASSO and interpret them as a sign-consistency condition.
Subdifferential of at is .
Subgradient optimality
.
Componentwise
For with : . For with : .
Interpretation
Inactive coordinates must have correlation with the residual bounded by .
ex-ch13-11
MediumA sensing matrix with , is drawn i.i.d.\ . Estimate the maximum sparsity for which RIP-based recovery is likely.
Use .
Solve for $s$
With (typical constant): . Trying : . So is feasible.
Practical answer
likely recoverable; unlikely.
ex-ch13-12
MediumShow that the proximal operator of is componentwise soft thresholding.
Minimize componentwise.
Scalar problem
. Subgradient condition: .
Cases
If : . If : . Else . This is .
ex-ch13-13
HardProve that if , then every -sparse is the unique -sparse solution of .
is the smallest number of linearly dependent columns.
Contrapositive
Suppose distinct -sparse give the same . Then is -sparse, nonzero, and , i.e.\ columns of on are dependent.
Conclude
This contradicts , so .
ex-ch13-14
HardDerive the Welch bound: for a unit-norm dictionary , .
Compute two ways.
Frobenius identity
.
Use rank bound
Since is of rank , .
Combine
(bounding the average off-diagonal by ). Solving for : .
ex-ch13-15
HardShow that if , BPDN's error is bounded by for a constant depending only on .
Use the cone condition and the RIP-based restricted eigenvalue property.
Cone condition
Let . Feasibility of both and gives . Optimality of gives where .
Norm compression
Let be the top- entries of , the next, etc. A standard bound gives .
Apply RIP
. The condition makes the coefficient positive; solving yields .