Research2026-04-22
VLA Foundry: A Unified Framework for Training Vision-Language-Action Models
Source: Arxiv CS.AI
arXiv:2604.19728v1 Announce Type: cross Abstract: We present VLA Foundry, an open-source framework that unifies LLM, VLM, and VLA training in a single codebase. Most open-source VLA efforts specialize on the action training stage, often stitching together incompatible pretraining pipelines. VLA...
arxivpapersvision