Google TurboQuant: 6x Less Memory, 8x Faster — The KV Cache Revolution
Every time you chat with a large language model, it’s quietly doing something expensive in the background: keeping a running record of everything you’ve said. That record — the key-value (KV) cache — is what allows the model to stay coherent across a long conversation. But as conversations grow, so does the cache, consuming GPU memory, slowing responses, and driving up inference costs.
Google Research just published TurboQuant, a compression algorithm that shrinks that cache by more than 6x — with zero accuracy loss — and speeds up attention computation by up to 8x on Nvidia H100s. No retraining required.
The paper is being presented at ICLR 2026 in April.
What Is the KV Cache, and Why Does It Matter?
When an LLM generates text, it computes attention over all previous tokens — checking how each new word relates to everything said before. To avoid recomputing this for every token, models cache the key and value matrices from each attention layer. That cache is the KV cache.
It’s fast, but it’s heavy. For a long conversation or a document with thousands of tokens, the KV cache can consume gigabytes of GPU memory per user session. At scale — serving thousands of concurrent users — this becomes a serious bottleneck. More memory pressure means fewer concurrent requests, slower throughput, and higher cost per token.
Shrinking the KV cache is one of the highest-leverage optimizations in LLM inference today.
How TurboQuant Works
Traditional vector quantization — representing high-dimensional vectors with fewer bits — is a well-understood compression technique. The catch: most quantization methods require storing quantization constants (the reference values used to decode compressed data) at full precision. That overhead can cost 1–2 extra bits per value, partially defeating the purpose.
TurboQuant eliminates that overhead through two steps:
1. PolarQuant — Geometry-aware compression Before quantizing, TurboQuant randomly rotates the data vectors. This geometric transformation equalizes the distribution of values across dimensions, making them easier to compress uniformly without per-block constants. The rotation is cheap to compute and, critically, requires no stored overhead at inference time.
2. Quantized Johnson-Lindenstrauss (QJL) — Overhead-free bit packing TurboQuant then applies QJL, a quantization scheme derived from the classical Johnson-Lindenstrauss lemma (which guarantees that random projections preserve distances). QJL achieves true 3-bit compression with zero stored constants — no metadata, no overhead.
The result: 3-bit KV cache with 6x memory reduction and no measurable accuracy loss.
In Google’s tests, TurboQuant scored perfectly on RULER — a benchmark specifically designed to stress-test long-context retrieval by burying a key detail in a large amount of text. This is exactly the scenario where KV cache quality matters most.
On Nvidia H100s, the attention computation speedup reached 8x versus standard 16-bit methods — no architectural changes, no additional runtime cost.
TurboQuant also outperformed rival methods on vector search tasks (think: semantic search, nearest-neighbor retrieval), which use similar high-dimensional compression techniques.
What This Means in Practice
The implications cascade in several directions.
Longer contexts, same hardware. A 6x memory reduction means you can serve 6x longer conversations on the same GPU, or serve 6x more users concurrently. For providers charging by token, that’s a direct cost reduction.
Wall Street felt it. Despite TurboQuant’s paper first appearing on arXiv in April 2025, the official Google Research blog post this week triggered a 3–5% selloff in AI memory companies. One paper won’t crater HBM demand — LLMs are still hungry — but the market is beginning to price in a world where smarter software compresses away some of the premium that AI commands on memory.
No retraining. This is the practical detail that makes TurboQuant immediately deployable. You don’t need to fine-tune or re-run pretraining. Drop it into your inference stack and get the compression.
The Bigger Picture: A Race to Run AI on Less
TurboQuant is one piece of a larger trend: the push to run capable AI on radically constrained hardware.
A related line of work — MatMul-Free Language Models from UC Santa Cruz and collaborators — takes a different but complementary approach. Rather than compressing the KV cache, it eliminates matrix multiplication from the model architecture entirely, replacing it with ternary weights and bitwise operations. Their models run on a custom FPGA chip that fits on a desk, consuming a fraction of the power of a GPU while matching transformer performance at up to 2.7B parameters.
Where TurboQuant says “compress the memory the model needs at runtime”, MatMul-free says “redesign the model so it barely needs memory at all”. Together, they represent two converging paths toward the same destination: LLMs that run cheaply, at scale, on smaller hardware.
The implications for edge deployment — on-device AI, embedded medical systems, inference on laptops without cloud calls — are significant.
The Technical Details Worth Knowing
If you’re building with this:
- Paper: TurboQuant on arXiv (2504.19874)
- Google blog post: research.google/blog/turboquant
- Safe to apply to: Any transformer-based LLM with standard attention; no architecture changes needed
- Tested on: Gemma and other Google models; generalizes to standard transformer architectures
- Compression level: 3-bit KV cache (down from typical 16-bit), 6x+ memory reduction
- Speedup: Up to 8x attention computation on H100; mileage varies by model and sequence length
- Companion methods: PolarQuant (AISTATS 2026) and QJL are published separately and can be used independently
Where This Is Headed
The KV cache is increasingly the bottleneck in LLM deployment — not the model weights themselves. Most models can already be quantized to 4-bit weights with minimal accuracy loss. The KV cache has been harder to crack cleanly, because errors there propagate directly into generation quality.
TurboQuant’s zero-accuracy-loss result at 3 bits, confirmed on RULER’s demanding long-context retrieval tasks, is a meaningful advance. If it holds up in production deployments across diverse workloads, expect to see it integrated into inference frameworks like vLLM, TGI, and TensorRT-LLM relatively quickly.
The era of treating GPU memory as an unlimited resource is ending. The algorithms are catching up.