Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2310.07147
Cited By
QFT: Quantized Full-parameter Tuning of LLMs with Affordable Resources
11 October 2023
Zhikai Li
Xiaoxuan Liu
Banghua Zhu
Zhen Dong
Qingyi Gu
Kurt Keutzer
MQ
Re-assign community
ArXiv
PDF
HTML
Papers citing
"QFT: Quantized Full-parameter Tuning of LLMs with Affordable Resources"
7 / 7 papers shown
Title
SAQ-SAM: Semantically-Aligned Quantization for Segment Anything Model
Jing Zhang
Z. Li
Qingyi Gu
MQ
VLM
48
0
0
09 Mar 2025
Stochastic Communication Avoidance for Recommendation Systems
Lutfi Eren Erdogan
Vijay Anand Raghava Kanakagiri
Kurt Keutzer
Zhen Dong
40
1
0
03 Nov 2024
Achieving Peak Performance for Large Language Models: A Systematic Review
Z. R. K. Rostam
Sándor Szénási
Gábor Kertész
26
3
0
07 Sep 2024
Exploring Quantization for Efficient Pre-Training of Transformer Language Models
Kamran Chitsaz
Quentin Fournier
Gonccalo Mordido
Sarath Chandar
MQ
36
3
0
16 Jul 2024
LLM Inference Unveiled: Survey and Roofline Model Insights
Zhihang Yuan
Yuzhang Shang
Yang Zhou
Zhen Dong
Zhe Zhou
...
Yong Jae Lee
Yan Yan
Beidi Chen
Guangyu Sun
Kurt Keutzer
37
77
0
26 Feb 2024
RepQuant: Towards Accurate Post-Training Quantization of Large Transformer Models via Scale Reparameterization
Zhikai Li
Xuewen Liu
Jing Zhang
Qingyi Gu
MQ
27
7
0
08 Feb 2024
I-ViT: Integer-only Quantization for Efficient Vision Transformer Inference
Zhikai Li
Qingyi Gu
MQ
41
94
0
04 Jul 2022
1