Scalability to Large Scenes

From Rooms to Cities

Most neural scene representation methods (NeRF, SDF, 3DGS) have been demonstrated on single rooms or small areas. Scaling to building-scale, campus-scale, or city-scale RF imaging presents fundamental challenges in memory, computation, and data collection. The right representation for a campus may not yet exist -- and finding it is where the next generation of researchers will make their contributions.

Definition:

Scalability Bottlenecks

Scaling RF imaging to large scenes faces three bottlenecks:

  1. Memory: a single room (10×10×310 \times 10 \times 3 m) at 5 cm resolution requires a 200×200×60200 \times 200 \times 60 grid (2.42.4M voxels). A campus (500×500×30500 \times 500 \times 30 m) requires 10,000×10,000×60010{,}000 \times 10{,}000 \times 600 (6×10106 \times 10^{10} voxels) -- infeasible for any explicit representation.

  2. Computation: rendering a single ray through a campus-scale scene requires evaluating thousands of Gaussians or MLP queries, scaling linearly with scene volume.

  3. Data collection: measuring RF propagation across a campus requires thousands of Tx-Rx locations, weeks of measurement time, and coordination across multiple buildings.

Computational Cost vs Scene Resolution\text{Computational Cost vs Scene Resolution}

Compares the memory and computation scaling of voxel grids, neural fields (NeRF), and hierarchical representations as scene size increases. Hierarchical methods maintain tractable cost by adapting resolution to the query location.

Parameters
50
5

Definition:

Hierarchical Scene Representations

Hierarchical methods address scalability by dividing the scene:

  • Spatial partitioning: octrees, block-NeRFs, or tiled 3DGS partition the scene into manageable blocks, each with its own representation.

  • Level of detail (LOD): coarse representation for distant regions, fine representation for nearby regions. For RF: coarse = path loss model, fine = neural field.

  • Streaming: load only the relevant blocks into GPU memory as the query region moves.

The composition rule for adjacent blocks must preserve physical consistency: signals crossing block boundaries must maintain phase continuity and amplitude consistency.

Theorem: Complexity of Hierarchical vs Flat Representations

For a scene of side length LL at resolution Δx\Delta x, a flat voxel grid requires O((L/Δx)3)\mathcal{O}((L/\Delta x)^3) memory. An octree with maximum depth D=log2(L/Δx)D = \log_2(L/\Delta x) and KK occupied leaf nodes requires O(K)\mathcal{O}(K) memory, where K(L/Δx)2K \leq (L/\Delta x)^2 for scenes with surface-like structure (walls, floors). The memory reduction is:

FlatOctree=(L/Δx)3(L/Δx)2=LΔx.\frac{\text{Flat}}{\text{Octree}} = \frac{(L/\Delta x)^3}{(L/\Delta x)^2} = \frac{L}{\Delta x}.

For a campus at 5 cm: reduction factor =500/0.05=10,000= 500/0.05 = 10{,}000.

Indoor scenes are mostly empty space. An octree only stores non-empty regions, and surfaces are 2D manifolds embedded in 3D space. The saving is proportional to the scene depth divided by the resolution.

Example: Campus-Scale Digital Twin Design

Design a hierarchical RF digital twin for a 500×500500 \times 500 m university campus with 20 buildings. The twin must support 5G channel prediction at 28 GHz within 32 GB GPU memory.

Open Problems in Scalable RF Imaging

  • Block boundary artifacts: how to seamlessly blend neural representations at block boundaries without phase or amplitude discontinuities?

  • Distributed training: can multiple base stations collaboratively build a campus-scale digital twin from local measurements, without sharing raw data (privacy)?

  • Incremental mapping: how to extend the map as new areas are explored, without retraining the entire representation?

  • Compression: learned scene representations must be compressed for storage and transmission. What are the rate-distortion limits for RF scene compression?

  • Multi-frequency consistency: a campus has sub-6 GHz macro cells and mmWave small cells. Can a single representation serve both frequency bands with shared geometry but frequency-dependent reflectivities?

🔧Engineering Note

GPU Memory Budget for Large-Scale RF Imaging

Modern GPUs provide 24--80 GB memory. A practical budget allocation for campus-scale RF digital twins:

  • Scene representation: 1--4 GB (Gaussians or neural field weights).
  • Forward model: 2--8 GB (sensing operator, precomputed kernels).
  • Optimiser state: 2--8 GB (Adam moments, gradient buffers).
  • Working memory: 4--16 GB (intermediate computations).

Total: 10--36 GB, fitting within a single A100 (80 GB) or requiring model parallelism across 2 consumer GPUs (2×\times24 GB). The forward model is often the largest component; the Kronecker structure from Chapter 7 is essential for keeping it tractable.

Common Mistake: Phase Discontinuity at Block Boundaries

Mistake:

Independently training adjacent blocks without enforcing continuity constraints. The result: phase jumps at block boundaries that corrupt channel predictions for signals crossing multiple blocks.

Correction:

Enforce overlap regions where adjacent blocks share measurements. Add a boundary consistency loss: Lboundary=γA(x)γB(x)2\mathcal{L}_{\mathrm{boundary}} = \|\boldsymbol{\gamma}_A(\mathbf{x}) - \boldsymbol{\gamma}_B(\mathbf{x})\|^2 for x\mathbf{x} in the overlap zone. Alternatively, use a global phase reference (e.g., from a reference anchor AP).

Hierarchical Scene Representation

A multi-resolution scene representation that allocates fine detail to regions of interest and coarse approximations to distant or empty regions. Enables scaling to campus and city-scale scenes within finite memory budgets.

Related: Primitive-Based Scene Representation

Key Takeaway

Memory, computation, and data collection are the three scalability bottlenecks. Hierarchical representations (octrees, block-NeRFs, tiled 3DGS) with level-of-detail reduce memory from O(N3)\mathcal{O}(N^3) to O(N2)\mathcal{O}(N^2) for surface-dominated scenes. Block boundary consistency, distributed training, and multi-frequency support remain significant open problems.