Research2026-05-14
GRIP-VLM: Group-Relative Importance Pruning for Efficient Vision-Language Models
Source: Arxiv CS.AI
arXiv:2605.13375v1 Announce Type: cross Abstract: In Vision-Language Models (VLMs), processing a massive number of visual tokens incurs prohibitive computational overhead. While recent training-aware pruning methods attempt to selectively discard redundant tokens, they largely rely on...
arxivpapersvision