BeClaude
Research2026-05-11

ForgeVLA: Federated Vision-Language-Action Learning without Language Annotations

Source: Arxiv CS.AI

arXiv:2605.07474v1 Announce Type: cross Abstract: Vision-Language-Action (VLA) models hold great promise for general-purpose robotic intelligence, yet scaling up such models is severely bottlenecked by the high cost of acquiring annotated training data. Fortunately, vision-equipped robots deployed...

arxivpapersvision