HUGSIM: A Real-Time, Photo-Realistic and Closed-Loop Simulator for Autonomous Driving

Your Image
1Zhejiang University, 2Huawei Noah's Ark Lab, 3University of Tübingen, 4Tübingen AI Center

Video

Overview

Teaser image demonstrating HUGSIM.

We present HUGSIM, a real-time, photo-realistic and closed-loop simulator for autonomous driving. HUGSIM enables the full closed simulation loop, dynamically updating the ego and actor positions and observations based on control commands. It incorporates 360-degree high-fidelity actor insertion and efficiently generates both normal and aggressive actor behaviors. Moreover, HUGSIM offers a comprehensive benchmark across more than 70 sequences from KITTI-360, Waymo Open Dataset, Nuscenes, and Pandaset, along with over 400 varying scenarios, providing a fair and realistic evaluation platform for existing autonomous driving algorithms.

HUGSIM Architecture

HUGSIM Architecture

Our algorithm takes as input posed images of a dynamic urban scene. We decompose the scene into static and dynamic 3D Gaussians, with the motion of dynamic vehicles being modeled via a unicycle model. The 3D Gaussians represent not only appearance but also semantic and flow information, allowing for rendering the RGB images, semantic labels, as well as optical flow through volume rendering.

Benchmark Evaluation

HUGSIM Benchmark

We evaluate UniAD, VAD and Latent-Transfuser on our HUGSIM benchmark.

Extrapolated Views Comparison


Benchmark Gallery

Original Scenarios
Edited Scenarios

This paper is an extension of HUGS: Holistic Urban 3D Scene Understanding via Gaussian Splatting

BibTeX


      @article{zhou2024hugsim,
        title={HUGSIM: A Real-Time, Photo-Realistic and Closed-Loop Simulator for Autonomous Driving},
        author={Zhou, Hongyu and Lin, Longzhong and Wang, Jiabao and Lu, Yichong and Bai, Dongfeng and Liu, Bingbing and Wang, Yue and Geiger, Andreas and Liao, Yiyi},
        journal={arXiv preprint arXiv:2412.01718},
        year={2024}
      }