Research2026-05-12
ALAM: Algebraically Consistent Latent Transitions for Vision-Language-Action Models
Source: Arxiv CS.AI
arXiv:2605.10819v1 Announce Type: cross Abstract: Vision-language-action (VLA) models remain constrained by the scarcity of action-labeled robot data, whereas action-free videos provide abundant evidence of how the physical world changes. Latent action models offer a promising way to extract such...
arxivpapersvision