Back to News
Research2026-04-17
KV Packet: Recomputation-Free Context-Independent KV Caching for LLMs
Source: Arxiv CS.AI
arXiv:2604.13226v1 Announce Type: cross Abstract: Large Language Models (LLMs) rely heavily on Key-Value (KV) caching to minimize inference latency. However, standard KV caches are context-dependent: reusing a cached document in a new context requires recomputing KV states to account for shifts in...
arxivpapers