Research2026-04-28
Vision-Language-Action in Robotics: A Survey of Datasets, Benchmarks, and Data Engines
Source: Arxiv CS.AI
arXiv:2604.23001v1 Announce Type: cross Abstract: Despite remarkable progress in Vision--Language--Action (VLA) models, a central bottleneck remains underexamined: the data infrastructure that underlies embodied learning. In this survey, we argue that future advances in VLA will depend less on...
arxivpapersbenchmarkvisionrobotics