LVLDrive: Enhancing Autonomous Driving with 3D Spatial Understanding

Published:Dec 30, 2025 16:35
1 min read
ArXiv

Analysis

This paper addresses a critical limitation of Vision-Language Models (VLMs) in autonomous driving: their reliance on 2D image cues for spatial reasoning. By integrating LiDAR data, the proposed LVLDrive framework aims to improve the accuracy and reliability of driving decisions. The use of a Gradual Fusion Q-Former to mitigate disruption to pre-trained VLMs and the development of a spatial-aware question-answering dataset are key contributions. The paper's focus on 3D metric data highlights a crucial direction for building trustworthy VLM-based autonomous systems.

Reference

LVLDrive achieves superior performance compared to vision-only counterparts across scene understanding, metric spatial perception, and reliable driving decision-making.