RadarGen: Automotive Radar Point Cloud Generation from Cameras

Reading time: 5 minute
...

📝 Original Info

  • Title: RadarGen: Automotive Radar Point Cloud Generation from Cameras
  • ArXiv ID: 2512.17897
  • Date: 2025-12-19
  • Authors: Researchers from original ArXiv paper

📝 Abstract

We present RadarGen, a diffusion model for synthesizing realistic automotive radar point clouds from multi-view camera imagery. RadarGen adapts efficient image-latent diffusion to the radar domain by representing radar measurements in bird's-eye-view form that encodes spatial structure together with radar cross section (RCS) and Doppler attributes. A lightweight recovery step reconstructs point clouds from the generated maps. To better align generation with the visual scene, RadarGen incorporates BEV-aligned depth, semantic, and motion cues extracted from pretrained foundation models, which guide the stochastic generation process toward physically plausible radar patterns. Conditioning on images makes the approach broadly compatible, in principle, with existing visual datasets and simulation frameworks, offering a scalable direction for multimodal generative simulation. Evaluations on large-scale driving data show that RadarGen captures characteristic radar measurement distributions and reduces the gap to perception models trained on real data, marking a step toward unified generative simulation across sensing modalities.

💡 Deep Analysis

Deep Dive into RadarGen: Automotive Radar Point Cloud Generation from Cameras.

We present RadarGen, a diffusion model for synthesizing realistic automotive radar point clouds from multi-view camera imagery. RadarGen adapts efficient image-latent diffusion to the radar domain by representing radar measurements in bird’s-eye-view form that encodes spatial structure together with radar cross section (RCS) and Doppler attributes. A lightweight recovery step reconstructs point clouds from the generated maps. To better align generation with the visual scene, RadarGen incorporates BEV-aligned depth, semantic, and motion cues extracted from pretrained foundation models, which guide the stochastic generation process toward physically plausible radar patterns. Conditioning on images makes the approach broadly compatible, in principle, with existing visual datasets and simulation frameworks, offering a scalable direction for multimodal generative simulation. Evaluations on large-scale driving data show that RadarGen captures characteristic radar measurement distributions an

📄 Full Content

RadarGen: Automotive Radar Point Cloud Generation from Cameras Tomer Borreda1 Fangqiang Ding1,2 Sanja Fidler3,4,5 Shengyu Huang3 Or Litany1,3 1 Technion 2 MIT 3 NVIDIA 4 University of Toronto 5 Vector Institute https://radargen.github.io/ Figure 1. Controllable radar synthesis from vision. (Top) Given multi-view camera images, RadarGen generates realistic radar point clouds that align with real-world radar statistics and can be consumed by downstream perception models. (Bottom) The generation is semantically consistent: modifying the input scene with an off-the-shelf image editing tool (e.g., replacing a distant car with a closer truck) updates the radar response, removing returns from newly occluded regions and reflecting the new object geometry. Abstract We present RadarGen, a diffusion model for synthesizing re- alistic automotive radar point clouds from multi-view cam- era imagery. RadarGen adapts efficient image-latent diffu- sion to the radar domain by representing radar measure- ments in bird’s-eye-view form that encodes spatial struc- ture together with radar cross section (RCS) and Doppler attributes. A lightweight recovery step reconstructs point clouds from the generated maps. To better align generation with the visual scene, RadarGen incorporates BEV-aligned depth, semantic, and motion cues extracted from pretrained foundation models, which guide the stochastic generation process toward physically plausible radar patterns. Con- ditioning on images makes the approach broadly compat- ible, in principle, with existing visual datasets and simu- lation frameworks, offering a scalable direction for multi- modal generative simulation. Evaluations on large-scale driving data show that RadarGen captures characteristic radar measurement distributions and reduces the gap to perception models trained on real data, marking a step to- ward unified generative simulation across sensing modali- ties. 1 arXiv:2512.17897v1 [cs.CV] 19 Dec 2025 1. Introduction Recent advances in neural and generative simulation have made it increasingly practical to synthesize photorealistic data at scale for autonomous driving. By reconstructing real scenes with neural fields or generating entirely new ones us- ing video diffusion models, these systems can produce di- verse and controllable environments that closely mimic real sensor observations. This capability enables large scale res- imulation of traffic, lighting, and weather conditions with- out costly rerecording or manual setup [43, 51, 62]. De- spite this rapid progress, most neural simulators remain lim- ited to the visual domain, focusing on the generation of RGB imagery and video. Recent efforts have begun ex- tending these ideas to LiDAR, demonstrating controllable three-dimensional point cloud generation from camera in- puts [57, 60, 90]. Radar, however, remains an open fron- tier. Although it is already ubiquitous in production vehi- cles, providing low cost, lightweight, and weather resilient perception, it has received far less attention within the gen- erative modeling community. This imbalance limits the fi- delity of current neural simulators, which cannot reproduce radar’s distinctive sensing characteristics, including signal sparsity, radar cross section (RCS), and Doppler. Generating radar data poses unique challenges. Radar measurements exhibit strong stochasticity due to multipath reflections, interference, and material-dependent scattering that vary with scene geometry. Operating at longer wave- lengths, radar interacts with surfaces and internal structures beyond what cameras or LiDAR perceive, making it highly complementary yet difficult to model from vision alone. A further challenge lies in the nature of available radar data. In most large scale driving datasets, radar is provided only after proprietary signal processing that converts raw radio frequency waveforms into sparse point clouds with RCS and Doppler values. This processing chain, which includes range Doppler transforms, beamforming, and detection al- gorithms such as constant false alarm rate (CFAR), is closed and lossy, discarding phase and other fine grained signal information. In practice, storing raw radar signals is ex- tremely memory intensive, so even commercial survey ve- hicles often record only processed point clouds. As a result, point clouds remain the practical representation of radar data for large scale learning. To address these challenges, we propose RadarGen, a generative framework for synthesizing automotive radar point clouds directly from camera imagery. RadarGen learns a distribution over radar observations conditioned on the visual scene, producing diverse and semantically consis- tent measurements rather than a single deterministic predic- tion, reflecting the inherent stochasticity of real radar sig- nals. Conditioning on images allows RadarGen to leverage existing visual data and simulators, providing a scalable and modular way to enrich them with realistic radar s

…(Full text truncated)…

📸 Image Gallery

fig_method_radar_maps.png fig_recovery.webp fig_teaser.png

Reference

This content is AI-processed based on ArXiv data.

Start searching

Enter keywords to search articles

↑↓
ESC
⌘K Shortcut