Search:
Match:
1 results

Analysis

This paper addresses a critical limitation of Vision-Language Models (VLMs) in autonomous driving: their reliance on 2D image cues for spatial reasoning. By integrating LiDAR data, the proposed LVLDrive framework aims to improve the accuracy and reliability of driving decisions. The use of a Gradual Fusion Q-Former to mitigate disruption to pre-trained VLMs and the development of a spatial-aware question-answering dataset are key contributions. The paper's focus on 3D metric data highlights a crucial direction for building trustworthy VLM-based autonomous systems.
Reference

LVLDrive achieves superior performance compared to vision-only counterparts across scene understanding, metric spatial perception, and reliable driving decision-making.