ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2503.01297
55
1

Regularization-based Framework for Quantization-, Fault- and Variability-Aware Training

3 March 2025
Anmol Biswas
Raghav Singhal
Sivakumar Elangovan
Shreyas Sabnis
U. Ganguly
    MQ
ArXivPDFHTML
Abstract

Efficient inference is critical for deploying deep learning models on edge AI devices. Low-bit quantization (e.g., 3- and 4-bit) with fixed-point arithmetic improves efficiency, while low-power memory technologies like analog nonvolatile memory enable further gains. However, these methods introduce non-ideal hardware behavior, including bit faults and device-to-device variability. We propose a regularization-based quantization-aware training (QAT) framework that supports fixed, learnable step-size, and learnable non-uniform quantization, achieving competitive results on CIFAR-10 and ImageNet. Our method also extends to Spiking Neural Networks (SNNs), demonstrating strong performance on 4-bit networks on CIFAR10-DVS and N-Caltech 101. Beyond quantization, our framework enables fault and variability-aware fine-tuning, mitigating stuck-at faults (fixed weight bits) and device resistance variability. Compared to prior fault-aware training, our approach significantly improves performance recovery under upto 20% bit-fault rate and 40% device-to-device variability. Our results establish a generalizable framework for quantization and robustness-aware training, enhancing efficiency and reliability in low-power, non-ideal hardware.

View on arXiv
@article{biswas2025_2503.01297,
  title={ Regularization-based Framework for Quantization-, Fault- and Variability-Aware Training },
  author={ Anmol Biswas and Raghav Singhal and Sivakumar Elangovan and Shreyas Sabnis and Udayan Ganguly },
  journal={arXiv preprint arXiv:2503.01297},
  year={ 2025 }
}
Comments on this paper