Research2026-04-24
SparKV: Overhead-Aware KV Cache Loading for Efficient On-Device LLM Inference
Source: Arxiv CS.AI
arXiv:2604.21231v1 Announce Type: cross Abstract: Efficient inference for on-device Large Language Models (LLMs) remains challenging due to limited hardware resources and the high cost of the prefill stage, which processes the full input context to construct Key-Value (KV) caches. We present...
arxivpapers