Exercises
ex26-01-gaussian-psd
EasyShow that the covariance parameterisation always produces a positive semi-definite matrix, regardless of and diagonal .
Consider for arbitrary .
Expand the quadratic form
$
Conclude
Since for any vector , the matrix is PSD for any rotation and any scale . It is strictly positive definite when all diagonal entries of are nonzero.
ex26-02-alpha-compositing
EasyConsider three Gaussians with opacities , , and features , , , all evaluated at a pixel where for all (the pixel is at the centre of each Gaussian). Compute the rendered value using alpha compositing.
Apply the front-to-back compositing formula.
Compute for each .
Compute transmittances
- (no preceding Gaussians)
Compute the rendered value
\blacksquare$
ex26-03-quaternion
EasyA 3D Gaussian has its rotation stored as the unit quaternion . What rotation matrix does this correspond to? What happens to the Gaussian's shape?
The identity quaternion corresponds to zero rotation.
Convert quaternion to rotation matrix
The quaternion gives (the identity matrix), since:
Interpret
With , the covariance becomes . The Gaussian is axis-aligned with semi-axes along the coordinate axes. No rotation is applied.
ex26-04-db-loss
EasyA measurement location has true received power dBm. Two models predict dBm and dBm. Compute the MSE loss in both dB scale and linear scale for each prediction, and explain why dB-scale loss is preferred.
Convert dBm to mW: .
dB-scale loss
- Model 1: dB
- Model 2: dB
Both predictions have equal error in dB scale.
Linear-scale loss
Converting: mW, mW, mW.
- Model 1:
- Model 2:
Model 2 has smaller loss in linear scale, despite having the same absolute error in dB. Linear-scale loss is dominated by high-power regions, ignoring errors at low power.
ex26-05-2d-projection
MediumDerive the 2D projected covariance of a 3D Gaussian with mean and covariance under a perspective camera with projection matrix . Show that the projection of an anisotropic 3D Gaussian is an anisotropic 2D Gaussian (ellipse) on the image plane.
Use the Jacobian of the projection at the Gaussian centre.
The projected distribution is obtained by marginalising over depth.
Linearise the projection
Let be the perspective projection. The Jacobian at is .
Transform the covariance
Under the linear approximation (EWA splatting), the projected covariance is:
where is the world-to-camera rotation.
Result is a 2D Gaussian
Since is PSD and is a matrix with rank 2 (for non-degenerate cameras), is a PSD matrix. The projected density is therefore a 2D Gaussian with elliptical level sets, centred at .
ex26-06-sh-isotropic
MediumShow that a spherical harmonic expansion truncated at order yields an isotropic (omnidirectional) scatterer, and compute the minimum order needed to represent a specular reflector with beamwidth .
is constant over the sphere.
The angular resolution of degree- SH is .
L=0 case
For : , which is constant over all directions. This is an omnidirectional (isotropic) scatterer.
Minimum L for specular pattern
The angular resolution of spherical harmonics of degree is approximately . To represent a specular reflection with beamwidth , we need:
For example, a specular reflector with beamwidth requires . In practice, -- suffices for typical RF scatterers because RF scattering patterns are broader than optical ones.
ex26-07-densification
MediumIn the 3DGS adaptive density control, a Gaussian is cloned when the positional gradient exceeds a threshold: . Explain geometrically what a large positional gradient indicates about the reconstruction quality at that location, and why cloning (adding a new Gaussian nearby) is the appropriate response.
The positional gradient points in the direction that would reduce the loss most.
A large gradient means the current Gaussian position is suboptimal.
Interpret the gradient
The positional gradient indicates the direction in which moving the Gaussian centre would most reduce the rendering loss. A large magnitude means the Gaussian is being "pulled" strongly --- it needs to cover more area than its current position and shape allow.
Why cloning helps
If a single Gaussian cannot adequately represent the local scene structure (e.g., a corner, an edge, or a transition between materials), duplicating it allows two Gaussians to share the responsibility. One stays near the current position; the clone shifts toward the gradient direction. After further optimisation, the two Gaussians settle into positions that jointly cover the under-reconstructed region.
Splitting vs cloning
Cloning is used for small Gaussians in under-represented regions; splitting is used for large Gaussians covering too much area. The distinction is controlled by the scale threshold .
ex26-08-measurement-density
MediumFor an RF-3DGS reconstruction at GHz, compute: (a) the wavelength , (b) the recommended measurement spacing (), (c) the number of measurements needed for a m room. Discuss practical feasibility.
where m/s.
Compute wavelength
$
Measurement spacing
Recommended spacing: cm.
Number of measurements
For a m area at 2.1 cm spacing: measurements.
Feasibility
Over 200,000 measurements is impractical for manual or even robotic data collection. This highlights why visual priors (RFCanvas) or coarser sampling with interpolation are essential at mmWave frequencies. Alternatively, phased-array beam scanning can acquire directional measurements more efficiently.
ex26-09-visual-prior
MediumIn the RFCanvas framework, the geometric parameters are frozen during RF fine-tuning. Compute the ratio of frozen to total parameters for a scene with Gaussians and SH order .
Geometric parameters per Gaussian: 3 (position) + 4 (quaternion) + 3 (scale) = 10.
RF parameters per Gaussian: 1 (opacity) + (SH) = ?
Count parameters
- Geometric per Gaussian:
- RF per Gaussian:
- Total per Gaussian:
- Total scene:
- Frozen (geometric):
Ratio
Frozen/total .
Half the parameters are frozen, meaning the RF optimisation operates in a 10,000-dimensional subspace. With 20 measurements, this is still severely under-determined (), which is why additional regularisation (smoothness, material-based priors) is important.
ex26-10-gradient-splatting
HardDerive the gradient of the rendering loss with respect to the opacity of the -th Gaussian, where is the alpha-compositing output. Show that the gradient depends on all Gaussians closer to the camera through the transmittance .
Write and note that .
The gradient has a direct term (from the -th summand) and indirect terms (from for ).
Direct contribution
The -th Gaussian contributes to . The direct gradient is:
Indirect contributions
For , the transmittance depends on : . This gives:
Total gradient
\partial\mathcal{L}/\partial\alpha_k = 2(\hat{C} - C_{\text{target}}) \cdot \partial\hat{C}/\partial\alpha_kT_k\blacksquare$
ex26-11-coherent-radar
HardShow that for two Gaussians at ranges and with (within the same range cell), the coherent radar return is:
and that the incoherent approximation can differ by up to a factor of 4 (6 dB).
Write and expand .
Coherent sum
$
Extreme cases
- Constructive: .
- Destructive: .
- Incoherent: .
For : constructive gives , incoherent gives , ratio (3 dB). For the full range: constructive/destructive ratio is (cancelled completely).
ex26-12-kronecker-gaussian
HardShow that when the sensing operator has Kronecker structure , the Gaussian splatting forward pass can be decomposed into separate spatial and frequency splatting operations, reducing the computational cost from to .
The Kronecker structure means the measurement at subcarrier and spatial sample factorises.
Each Gaussian contributes independently along spatial and frequency dimensions.
Kronecker forward model
The measurement at subcarrier and spatial position is:
Gaussian representation
Representing via Gaussians: , the measurement becomes:
Cost reduction
The frequency splatting requires operations and the spatial splatting , compared to for the joint operation. The total cost is rather than .
ex26-13-pruning
HardConsider a 3DGS scene with Gaussians after training. Suppose we prune all Gaussians with . Derive an upper bound on the rendering error introduced by pruning, in terms of and the maximum feature magnitude .
Each pruned Gaussian contributes at most to any pixel.
At most Gaussians can be pruned.
Bound per-Gaussian contribution
A pruned Gaussian with contributes at most to any pixel (since and ).
Bound total error
If Gaussians are pruned, the worst-case pixel error is:
For , , : error , which is a loose bound. In practice, the transmittance ensures that deeply occluded pruned Gaussians contribute negligibly, making the actual error much smaller.
ex26-14-convergence
ChallengeThe 3DGS optimisation minimises over the Gaussian parameters . This is a non-convex optimisation due to the sorting, alpha-compositing, and projection operations. Identify three specific sources of non-convexity and discuss under what conditions gradient descent is likely to find a good (if not global) minimum.
Consider the depth ordering (sorting), the product in , and the quaternion parameterisation.
Think about local vs global minima and symmetries of the representation.
Source 1: Depth sorting
The alpha-compositing depends on the depth order of Gaussians. Swapping the order of two Gaussians with similar depth creates a discontinuity in the rendering function. This makes the loss landscape non-smooth (not just non-convex). In practice, the sorting is treated as fixed during each gradient step and updated periodically.
Source 2: Transmittance product
The transmittance is a product of terms, each depending on different parameters. This creates coupled non-linear interactions between Gaussians --- the gradient of one Gaussian's opacity depends on all preceding Gaussians' opacities.
Source 3: Rotation parameterisation
The quaternion-to-rotation conversion is non-linear, and the rotation group SO(3) is non-convex. Additionally, antipodal quaternions and represent the same rotation, creating symmetries in the loss landscape.
Why gradient descent works
Despite non-convexity, 3DGS optimisation works well because: (1) The initialisation from SfM or a grid provides a warm start near a good basin. (2) Adaptive density control adds/removes parameters, exploring the loss landscape more broadly than fixed-parameter optimisation. (3) The per-Gaussian parameters are largely decoupled in non-overlapping regions.
ex26-15-fundamental-limits
ChallengeConsider the problem of recovering Gaussian parameters from RF power measurements. Using information-theoretic arguments, derive a necessary condition on for identifiability of the Gaussian scene, and compare with the achievability result from compressed sensing (RIP-based recovery).
Each Gaussian has parameters (--). The total scene has parameters.
Each measurement provides at most 1 real-valued constraint.
For identifiability, is necessary. Is it sufficient?
Necessary condition
The Gaussian scene has parameters. Each measurement provides one scalar equation. For the system to be determined, we need (parameter counting).
Structured observation
However, the rendering equation is non-linear in , so the equations are not independent linear constraints. The effective number of independent constraints depends on the Jacobian rank . For well-separated Gaussians, the Jacobian has full rank and suffices. For overlapping Gaussians, degeneracies reduce the rank.
Comparison with compressed sensing
In compressed sensing with a linear model , recovery of an -sparse signal in requires measurements under RIP. For Gaussian splatting with scatterers (each modelled as one Gaussian with parameters), the analogous requirement is --- the per-scatterer cost increases by the factor due to the additional shape parameters.
Practical implication
For scatterers in a -voxel space with : compressed sensing needs measurements; Gaussian splatting needs in the worst case. The visual prior (RFCanvas) dramatically reduces this by fixing the geometric parameters.