3D Gaussian Driving World Model for Unified Scene Understanding and Multi-Modal Generation

Published:Dec 29, 2025 03:40
1 min read
ArXiv

Analysis

This paper introduces a novel Driving World Model (DWM) that leverages 3D Gaussian scene representation to improve scene understanding and multi-modal generation in driving environments. The key innovation lies in aligning textual information directly with the 3D scene by embedding linguistic features into Gaussian primitives, enabling better context and reasoning. The paper addresses limitations of existing DWMs by incorporating 3D scene understanding, multi-modal generation, and contextual enrichment. The use of a task-aware language-guided sampling strategy and a dual-condition multi-modal generation model further enhances the framework's capabilities. The authors validate their approach with state-of-the-art results on nuScenes and NuInteract datasets, and plan to release their code, making it a valuable contribution to the field.

Reference

Our approach directly aligns textual information with the 3D scene by embedding rich linguistic features into each Gaussian primitive, thereby achieving early modality alignment.