Datasets for RF Imaging Research
Data Is the Foundation
Every RF imaging algorithm must be evaluated on data. Real measured data is scarce and expensive to collect; synthetic data is abundant but risks the inverse crime. This section surveys the standard datasets available to the community, describes how to generate synthetic data responsibly, and formalises the inverse crime --- the most common methodological sin in computational imaging.
Definition: 3D Shape Datasets for Synthetic RF Imaging
3D Shape Datasets for Synthetic RF Imaging
ShapeNet: a large repository of 3D CAD models ( models in 55 categories). Used to generate diverse target geometries for training learned reconstruction methods. Each model can be voxelised onto the imaging grid to produce a ground-truth reflectivity map.
THuman: a collection of 3D human body meshes from body scanning. Used for human-body RF imaging (e.g., through-wall sensing, vital sign monitoring) where the target has realistic human anatomy and posture.
MPEG-7 Shape Dataset: 2D shape silhouettes used as targets for 2D RF imaging experiments. Simpler than 3D models but useful for rapid prototyping and algorithm comparison.
Definition: Measurement Datasets for RF Imaging
Measurement Datasets for RF Imaging
Automotive radar datasets:
- nuScenes: 1000 driving scenes with 5 radar sensors, camera, and LiDAR. 3D bounding boxes for 23 object classes.
- RadarScenes: 158 sequences, 4 hours of driving data with radar point clouds and annotations.
Indoor imaging datasets:
- DeepMIMO: simulated massive MIMO channel dataset from ray tracing, covering multiple indoor and outdoor scenarios at 2.4, 28, and 60 GHz.
- DeepSense 6G: multi-modal sensing dataset for 6G research, including radar, camera, LiDAR, and GPS data.
SAR datasets:
- MSTAR: Moving and Stationary Target Recognition dataset. SAR images of military vehicles at various aspect angles.
- SEN12MS: multi-modal satellite dataset (SAR + optical) for land use classification.
Most datasets provide processed images or point clouds rather than raw radar measurements. For algorithm development at the signal processing level, raw ADC data or channel measurements are needed --- these are less commonly available.
Definition: Synthetic Data Generation Pipeline
Synthetic Data Generation Pipeline
When real data is unavailable or insufficient, simulation provides controlled, labelled data. A synthetic data pipeline has four stages:
-
Scene sampling: draw a random scene from a distribution (e.g., sample a ShapeNet model, place it at random position/orientation, set reflectivity parameters).
-
Forward model: compute synthetic measurements using the chosen physics (Born model, ray tracing, or full-wave).
-
Noise and impairments: add receiver noise (), phase noise, clutter, and hardware impairments.
-
Ground truth pairing: store both the measurement and the scene as a (input, target) pair for supervised training.
The critical question: does the forward model used for data generation match the one used for reconstruction? If yes, you have committed the inverse crime.
Definition: The Inverse Crime
The Inverse Crime
The inverse crime is committed when the same forward model generates the synthetic data and reconstructs from it:
When , the reconstruction problem reduces to denoising: the model mismatch error is zero, and the only error source is noise. This makes any reconstruction method look artificially good.
The crime is aggravated when:
- The same discretisation grid is used for generation and reconstruction;
- Point scatterers are placed exactly on grid points;
- The same random seed generates training and test data.
Theorem: Error Inflation from the Inverse Crime
Let be the true physics and the approximate model used for both simulation and reconstruction. The reconstruction error with the crime is:
while the honest error (different models) is:
The model mismatch term is absent in the crime, leading to optimistic error bounds.
Proof
With the crime: . The error is .
With model mismatch: . The additional term quantifies the model mismatch contribution.
Inverse Crime Demonstration
Demonstrate the inverse crime for a 1D imaging scenario. The plot shows: (left) the ground truth scene, (centre) the reconstruction via least squares, and (right) the error map.
Same grid (crime): toggle ON to use identical grids for simulation and reconstruction. The reconstruction is nearly perfect.
Different grids (honest): toggle OFF to use a finer grid (256 points) for simulation and a coarser grid (64 points) for reconstruction. Model mismatch causes visible artifacts, revealing the true performance.
Parameters
Definition: Strategies for Avoiding the Inverse Crime
Strategies for Avoiding the Inverse Crime
To avoid the inverse crime, ensure the forward model used for data generation differs meaningfully from the reconstruction model:
-
Different discretisations: generate data on a fine grid () and reconstruct on a coarser grid. Rule of thumb: .
-
Different physics: generate with ray tracing or FDTD; reconstruct with the Born approximation.
-
Off-grid scatterers: place point targets at positions that do not coincide with the reconstruction grid.
-
Model mismatch injection: add calibration errors, timing offsets, and mutual coupling to the simulated data but not to the reconstruction model.
-
Real data validation: always validate on measured data when available.
Example: Detecting the Inverse Crime in a Paper
A paper claims 45 dB PSNR for CS reconstruction of a radar image using LASSO with optimised on a validation set. The data is generated using a point-scatterer model with 20 targets on a grid. The same model is used for LASSO's matrix. Is the inverse crime present?
Crime identification
Yes, this is a textbook inverse crime: (same point-scatterer model). Furthermore, if the targets are on-grid, the sparsity assumption is exactly satisfied, giving LASSO an unfair advantage.
Expected honest performance
With model mismatch (e.g., ray tracing for forward, Born for reconstruction), PSNR typically drops by 10--20 dB. With off-grid targets, an additional 3--5 dB loss occurs due to basis mismatch. Honest estimate: 20--30 dB PSNR.
Recommendation
The paper should (1) use a different (higher-fidelity) forward model for data generation, (2) place targets off-grid, and (3) validate on measured data. The 45 dB claim is not credible for real-world deployment.
Common Mistake: The Inverse Crime Is Not Just About Grids
Mistake:
Believing that using a slightly different grid spacing avoids the inverse crime --- e.g., generating on a grid and reconstructing on .
Correction:
The grid difference must introduce meaningful model mismatch. A vs. grid produces near-zero mismatch. Use at least oversampling for generation (), and ideally use a different physics model (ray tracing vs. Born) or inject hardware impairments (phase noise, mutual coupling) that the reconstruction model does not know about.
Common Mistake: On-Grid Targets Inflate Sparse Recovery Performance
Mistake:
Placing point scatterers exactly on the reconstruction grid when evaluating sparse recovery algorithms (LASSO, OMP).
Correction:
On-grid targets perfectly match the dictionary, making the problem artificially easy. In reality, targets are continuous and the grid introduces basis mismatch. Always place targets at positions that do not coincide with the reconstruction grid (off-grid), or use super-resolution methods that explicitly model the continuous parameter.
Historical Note: The Name "Inverse Crime"
2000sThe term "inverse crime" was coined by Armand Wirgin in 2004, though the pitfall was well known in the inverse problems community for decades before. Colton and Kress formalised the concept in their classic textbook on inverse scattering theory. Despite widespread awareness, the inverse crime continues to appear in published RF imaging papers, often unintentionally, because the default simulation setup (generate and reconstruct with the same code) commits the crime by default.
Inverse Crime
The methodological error of using the same forward model for both synthetic data generation and reconstruction, producing artificially optimistic results by eliminating model mismatch.
ShapeNet
A large-scale 3D model repository containing over 51,000 CAD models in 55 categories, widely used for generating training data in computational imaging and computer vision research.
Quick Check
A paper reports 42 dB PSNR using a deep unrolling network trained and tested on data from the same Born-approximation forward model with on-grid targets. What is the most likely explanation for this high PSNR?
The deep unrolling architecture is extremely powerful
The inverse crime: same forward model and on-grid targets
The SNR is very high
The scene is very sparse
Correct. Eliminating model mismatch and basis mismatch produces artificially high PSNR.
Key Takeaway
Standard datasets (nuScenes, MSTAR, DeepMIMO) provide benchmarking data, but raw signal-level data remains scarce. Synthetic generation fills the gap but demands vigilance against the inverse crime: use different grids ( oversampling), different physics models, off-grid targets, and always validate on measured data. Papers reporting PSNR dB on purely simulated data should be scrutinised.