ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2026 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2502.04420
  4. Cited By
KVTuner: Sensitivity-Aware Layer-Wise Mixed-Precision KV Cache Quantization for Efficient and Nearly Lossless LLM Inference
v1v2v3v4v5 (latest)

KVTuner: Sensitivity-Aware Layer-Wise Mixed-Precision KV Cache Quantization for Efficient and Nearly Lossless LLM Inference

6 February 2025
Xianrui Li
Zeyu Xing
Yongqian Li
Linping Qu
Hui-Ling Zhen
Wulong Liu
Yiwu Yao
Sinno Jialin Pan
Mingxuan Yuan
    MQ
ArXiv (abs)PDFHTMLGithub (24★)

Papers citing "KVTuner: Sensitivity-Aware Layer-Wise Mixed-Precision KV Cache Quantization for Efficient and Nearly Lossless LLM Inference"

0 / 0 papers shown

No papers found

Page 1 of 0