BeClaude
Research2026-05-14

GRIP-VLM: Group-Relative Importance Pruning for Efficient Vision-Language Models

Source: Arxiv CS.AI

arXiv:2605.13375v1 Announce Type: cross Abstract: In Vision-Language Models (VLMs), processing a massive number of visual tokens incurs prohibitive computational overhead. While recent training-aware pruning methods attempt to selectively discard redundant tokens, they largely rely on...

arxivpapersvision