cmd2001 / KVTunerView on GitHub
[ICML2025] KVTuner: Sensitivity-Aware Layer-wise Mixed Precision KV Cache Quantization for Efficient and Nearly Lossless LLM Inference
26Jan 27, 2026Updated last month

Alternatives and similar repositories for KVTuner

Users that are interested in KVTuner are comparing it to the libraries listed below

Sorting:

Are these results useful?