Fast Algorithms for Structured Operators
The Computational Bottleneck
Every iterative reconstruction algorithm β ISTA, FISTA, ADMM, Chambolle--Pock β requires repeated evaluation of the forward operator and its adjoint . For a 2D imaging problem with measurements and voxels, a naive matrix-vector product costs floating-point operations. When and , this is manageable. But in 3D problems or dense multi-frequency configurations, and can each exceed , and storing the full matrix β let alone multiplying by it β becomes infeasible.
The point is that the sensing operator in RF imaging is not an arbitrary dense matrix. It has structure inherited from the physics: separable array geometries produce Kronecker products, uniform grids produce Fourier structure, and the combination of both yields algorithms that are orders of magnitude faster than the naive approach.
Definition: Kronecker Product
Kronecker Product
For and , the Kronecker product is defined by
Equivalently, is the block matrix whose -block is .
The Kronecker product appears naturally whenever the sensing geometry is separable β for example, when the transmit array, receive array, and frequency grid factor independently. This is the typical situation in our RF imaging system (The Kronecker Structure and Sensing Operator Properties).
Theorem: The Vec Trick for Kronecker Products
Let , , and . Then
Instead of forming the Kronecker product and multiplying by the -dimensional vectorized image, we can reshape the image into a matrix and multiply from both sides. The right-hand side involves two smaller matrix products, which is dramatically cheaper.
Column permutation argument
Write where is the -th column of . Applying :
Recognize the vectorized product
The -th column of is . Vectorizing this matrix yields exactly the expression above.
Complexity Reduction via the Vec Trick
The naive product costs operations and requires storing the full Kronecker product.
The factored computation costs:
- :
- Result :
- Total:
For our RF imaging system with and , the naive approach requires operations. The factored approach requires β a three-orders-of-magnitude reduction.
Storage is equally dramatic: the full matrix occupies million complex entries, while the two factors together need only entries.
Example: Kronecker Structure in the RF Imaging Sensing Operator
Consider a monostatic RF imaging system where the sensing operator for a single subcarrier has the form
where is the pilot matrix and contain steering vector entries.
For a uniform linear array imaging a rectangular grid, show that the core of the operator β the steering-vector part β has Kronecker structure, and identify the factors.
Identify the separable geometry
For a ULA with elements at positions imaging a 2D grid with coordinates , the steering vector for voxel has elements
In the far-field limit, this separates as
where depends on both and .
Kronecker factorization for 2D grid
When the grid is rectangular and the array is a UPA (uniform planar array) with elements, the steering matrix factors as
where and . Each factor is a partial DFT matrix.
Applying the vec trick
Reshape the image into . Then
which costs instead of .
Definition: The Fast Fourier Transform (FFT)
The Fast Fourier Transform (FFT)
The discrete Fourier transform (DFT) of is
The FFT computes in operations using the Cooley-Tukey divide-and-conquer factorization, compared to for the direct sum.
When the Kronecker factor is a DFT matrix (as occurs for uniform spatial sampling), the matrix-vector product reduces to an FFT. This is the case whenever the imaging grid has uniform spacing in the direction corresponding to .
In the Kronecker-factored computation , if both factors are DFT matrices, each column-wise or row-wise multiplication is an FFT. The total cost drops to .
Definition: Non-Uniform FFT (NUFFT)
Non-Uniform FFT (NUFFT)
The non-uniform FFT evaluates sums of the form
where the frequency points lie on a non-uniform grid, in operations.
The NUFFT proceeds in three steps:
- Gridding: Convolve the non-uniform samples onto an oversampled uniform grid using a compactly supported kernel (e.g., Kaiser-Bessel or Gaussian).
- FFT: Apply a standard FFT on the oversampled grid.
- Deapodization: Correct for the convolution kernel in the transform domain.
Accuracy depends on the oversampling factor (typically ) and the kernel width (typically 6--12 points). For imaging applications, relative errors below are standard.
The NUFFT is essential for RF imaging because the wavenumber samples determined by the array geometry and carrier frequencies rarely fall on a uniform grid. The NUFFT avoids the need to interpolate measurements onto a regular grid (which introduces additional approximation error).
Historical Note: The FFT: A Discovery Rediscovered
1805--1965The Cooley-Tukey FFT algorithm, published in 1965, reduced the DFT from to and is often cited as one of the ten most important algorithms of the twentieth century. But the idea was not new: Gauss had described an equivalent procedure in 1805 for interpolating asteroid orbits, and Runge published a similar scheme in 1903. The 1965 rediscovery by Cooley and Tukey was driven by Cold War signal processing needs β specifically, detecting nuclear tests from seismic data. The timing was perfect: digital computers had become fast enough to exploit the algorithm, and the applications were urgent enough to fund the research.
Theorem: Adjoint of a Kronecker-Structured Operator
If , then the adjoint (conjugate transpose) factors as
and the adjoint application satisfies
In particular, both the forward operator and the adjoint can be evaluated using the vec trick, with no additional memory or algorithmic complexity.
The adjoint of a Kronecker product is the Kronecker product of the adjoints β the factorization is preserved. This is essential because every proximal algorithm requires both (forward) and (adjoint) at every iteration.
Mixed-product property
For any matrices of compatible sizes, . Taking , and applying the transpose to both sides yields the result.
Adjoint vec trick
Kronecker-Factored Forward and Adjoint Operations
Complexity: for forward; same for adjointWhen and are DFT matrices, steps 2--3 reduce to batched FFTs, further lowering the cost to .
Kronecker Vec Trick: Speedup over Naive Product
Compare the wall-clock time of the naive dense matrix-vector product against the factored computation as the problem dimension grows. The speedup is most dramatic when the factor dimensions are balanced (, ).
Parameters
Maximum size of each Kronecker factor
Example: NUFFT vs DFT for Non-Uniform Wavenumber Samples
In a bistatic RF imaging configuration, the wavenumber samples are determined by the transmitter and receiver positions and do not lie on a uniform grid. Compare the direct evaluation
against the NUFFT approximation in terms of accuracy and computational cost for grid points and measurements.
Direct evaluation cost
Each of the outputs requires a sum over terms, giving total cost .
NUFFT cost
With oversampling, the NUFFT evaluates on a -point uniform grid via FFT (), then interpolates to the non-uniform points ( for kernel width ). Total: operations.
Accuracy comparison
For kernel width and oversampling , the NUFFT achieves relative error compared to the direct sum. This is more than sufficient for imaging applications where the measurement noise floor is typically at to dB.
NUFFT Approximation Error vs Kernel Width
Evaluate the relative error of the NUFFT approximation compared to direct DFT evaluation as a function of the interpolation kernel width . Demonstrates that -- suffices for imaging applications.
Parameters
Common Mistake: Kronecker Factor Ordering
Mistake:
Writing instead of and then applying the vec trick as .
Correction:
The Kronecker product is not commutative: in general. The vec trick maps the first factor to the right-multiplication and the second factor to the left-multiplication:
Swapping the factors produces a permuted result. Always verify by checking the dimensions: must be (rows match the second factor's columns).
Theorem: Spectral Properties of Kronecker Products
Let and have singular values and respectively. Then:
-
The singular values of are for all pairs .
-
.
-
, where denotes the condition number.
The condition number of the full sensing operator is the product of the condition numbers of the factors. If each factor is moderately ill-conditioned (say ), the Kronecker product can be severely ill-conditioned (). This multiplicative growth of ill-conditioning is a fundamental challenge in structured inverse problems.
SVD of the Kronecker product
Write and . By the mixed-product property,
Since Kronecker products of unitary matrices are unitary, this is a valid SVD. The diagonal entries of are all products .
Spectral norm and condition number
.
Similarly, . The ratio gives .
Quick Check
For and , the Kronecker product is a matrix of size:
.
Quick Check
When applying the vec trick to compute with and , what shape should be?
. The number of rows matches 's columns; the number of columns matches 's columns.
Definition: Multi-Factor Kronecker Products
Multi-Factor Kronecker Products
For three or more factors, the Kronecker product extends associatively:
The vec trick generalizes to tensor contractions. For the three-factor case (e.g., frequency Rx Tx in the RF imaging model), reshape into a 3D tensor and apply each factor along the corresponding mode:
where denotes the mode- product. The total cost is instead of .
In the RF imaging system of The Kronecker Structure and Sensing Operator Properties, the three factors correspond to the transmit steering (), receive steering (), and frequency phase () components.
Common Mistake: Near-Field Breaks Exact Kronecker Structure
Mistake:
Assuming that the sensing operator always has exact Kronecker structure, even when the array is in the near-field of the imaging region.
Correction:
The Kronecker factorization is exact only in the far-field regime, where the steering vector separates into azimuth and elevation components. In the near-field, the steering vector for each antenna depends on the full 3D distance , which does not factor.
For our system with Fresnel number (Section 5 of the code reference), near-field effects are non-negligible. In practice, one can:
- Use the Kronecker structure as a preconditioner (fast approximate inverse).
- Store and apply the exact operator, but use the Kronecker approximation for gradient computations where high accuracy is less critical.
- Apply a NUFFT-based approach that handles the non-uniform sampling directly.
Why This Matters: Kronecker Structure in MIMO Channel Estimation
The Kronecker structure exploited here for RF imaging is the same structure that appears in MIMO channel estimation. When the transmit and receive correlation matrices are separable (), the vectorized channel has a Kronecker-structured covariance. The LMMSE channel estimator then exploits this factorization in exactly the same way as our vec trick β the two-sided matrix multiplication replaces a large matrix inversion.
The insight that separable structure reduces computational cost from to in MIMO estimation (Kronecker Channel Models) is the same insight applied here to reduce the imaging operator cost.
Kronecker product
The tensor product of two matrices , formed by replacing each entry with the block . The result has dimensions .
Related: {{Ref:Gloss Vec Operator}}
Vec operator
The vectorization stacks the columns of into a single vector in . The inverse operation reshapes a vector back into a matrix.
Related: {{Ref:Gloss Kronecker Product}}
Non-uniform FFT (NUFFT)
An algorithm that evaluates Fourier sums at non-uniformly spaced frequency points in time, where is the number of source points and is the number of target points. Uses gridding, FFT, and deapodization.
Key Takeaway
The sensing operator in RF imaging inherits Kronecker structure from the separable array geometry. The vec trick converts a dense matrix-vector product into two smaller matrix multiplications, reducing cost by orders of magnitude. When the factors are DFT matrices (uniform grids), the FFT provides an additional logarithmic speedup. For non-uniform wavenumber sampling, the NUFFT achieves near-optimal complexity with controllable approximation error. Every iterative reconstruction algorithm benefits directly from these fast operator evaluations.