Research2026-05-07
VLMaxxing through FrameMogging Training-Free Anti-Recomputation for Video Vision-Language Models
Source: Arxiv CS.AI
arXiv:2605.03351v1 Announce Type: cross Abstract: Video vision-language models (VLMs) keep paying for visual state the stream already told us was stable. The factory wall did not move, but most VLM pipelines still hand the model dense RGB frames or a fresh prefix again. We study that waste as...
arxivpapersvision