ResearchTrend.AI
  • Papers
  • Communities
  • Organizations
  • Events
  • Blog
  • Pricing
  • Feedback
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2508.05571
66
0
v1v2v3 (latest)

Fairy±i\pm i±i: the First 2-bit Complex LLM with All Parameters in {±1,±i}\{\pm1, \pm i\}{±1,±i}

7 August 2025
Feiyu Wang
Guoan Wang
Yihao Zhang
S. Wang
Weitao Li
Bokai Huang
Shimao Chen
Z. L. Jiang
Rui Xu
Tong Yang
    MQ
ArXiv (abs)PDFHTMLGithub (101★)
Main:9 Pages
9 Figures
Bibliography:2 Pages
6 Tables
Appendix:4 Pages
Abstract

Quantization-Aware Training (QAT) integrates quantization into the training loop, enabling LLMs to learn robust low-bit representations, and is widely recognized as one of the most promising research directions. All current QAT research focuses on minimizing quantization error on full-precision models, where the full-precision accuracy acts as an upper bound (accuracy ceiling). No existing method has even attempted to surpass this ceiling. To break this ceiling, we propose a new paradigm: raising the ceiling (full-precision model), and then still quantizing it efficiently into 2 bits. We propose Fairy±i\pm i±i, the first 2-bit quantization framework for complex-valued LLMs. Specifically, our method leverages the representational advantages of the complex domain to boost full-precision accuracy. We map weights to the fourth roots of unity {±1,±i}\{\pm1, \pm i\}{±1,±i}, forming a perfectly symmetric and information-theoretically optimal 2-bit representation. Importantly, each quantized weight has either a zero real or imaginary part, enabling multiplication-free inference using only additions and element swaps. Experimental results show that Fairy±i\pm i±i outperforms the ceiling of existing 2-bit quantization approaches in terms of both PPL and downstream tasks, while maintaining strict storage and compute efficiency. This work opens a new direction for building highly accurate and practical LLMs under extremely low-bit constraints.

View on arXiv
Comments on this paper