BeClaude
Research2026-04-27

DVGT-2: Vision-Geometry-Action Model for Autonomous Driving at Scale

Source: Arxiv CS.AI

arXiv:2604.00813v3 Announce Type: replace-cross Abstract: End-to-end autonomous driving has evolved from the conventional paradigm based on sparse perception into vision-language-action (VLA) models, which focus on learning language descriptions as an auxiliary task to facilitate planning. In this...

arxivpapersvision