ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1910.04540
  4. Cited By
QPyTorch: A Low-Precision Arithmetic Simulation Framework

QPyTorch: A Low-Precision Arithmetic Simulation Framework

9 October 2019
Tianyi Zhang
Zhiqiu Lin
Guandao Yang
Christopher De Sa
    MQ
ArXivPDFHTML

Papers citing "QPyTorch: A Low-Precision Arithmetic Simulation Framework"

23 / 23 papers shown
Title
Pychop: Emulating Low-Precision Arithmetic in Numerical Methods and Neural Networks
Pychop: Emulating Low-Precision Arithmetic in Numerical Methods and Neural Networks
Erin Carson
Xinye Chen
49
0
0
10 Apr 2025
Scaling Laws for Floating Point Quantization Training
Scaling Laws for Floating Point Quantization Training
X. Sun
Shuaipeng Li
Ruobing Xie
Weidong Han
Kan Wu
...
Yangyu Tao
Zhanhui Kang
C. Xu
Di Wang
Jie Jiang
MQ
AIFin
60
0
0
05 Jan 2025
HDReason: Algorithm-Hardware Codesign for Hyperdimensional Knowledge
  Graph Reasoning
HDReason: Algorithm-Hardware Codesign for Hyperdimensional Knowledge Graph Reasoning
Hanning Chen
Yang Ni
Ali Zakeri
Zhuowen Zou
Sanggeon Yun
Fei Wen
Behnam Khaleghi
Narayan Srinivasa
Hugo Latapie
Mohsen Imani
31
1
0
09 Mar 2024
Trainable Fixed-Point Quantization for Deep Learning Acceleration on
  FPGAs
Trainable Fixed-Point Quantization for Deep Learning Acceleration on FPGAs
Dingyi Dai
Yichi Zhang
Jiahao Zhang
Zhanqiu Hu
Yaohui Cai
Qi Sun
Zhiru Zhang
MQ
54
5
0
31 Jan 2024
Towards Cheaper Inference in Deep Networks with Lower Bit-Width
  Accumulators
Towards Cheaper Inference in Deep Networks with Lower Bit-Width Accumulators
Yaniv Blumenfeld
Itay Hubara
Daniel Soudry
37
3
0
25 Jan 2024
ZeroQuant(4+2): Redefining LLMs Quantization with a New FP6-Centric
  Strategy for Diverse Generative Tasks
ZeroQuant(4+2): Redefining LLMs Quantization with a New FP6-Centric Strategy for Diverse Generative Tasks
Xiaoxia Wu
Haojun Xia
Stephen Youn
Zhen Zheng
Shiyang Chen
...
Reza Yazdani Aminabadi
Yuxiong He
Olatunji Ruwase
Leon Song
Zhewei Yao
66
8
0
14 Dec 2023
INT-FP-QSim: Mixed Precision and Formats For Large Language Models and
  Vision Transformers
INT-FP-QSim: Mixed Precision and Formats For Large Language Models and Vision Transformers
Lakshmi Nair
Mikhail Bernadskiy
Arulselvan Madhavan
Craig Chan
Ayon Basumallik
D. Bunandar
MQ
28
2
0
07 Jul 2023
Training with Mixed-Precision Floating-Point Assignments
Training with Mixed-Precision Floating-Point Assignments
Wonyeol Lee
Rahul Sharma
A. Aiken
MQ
19
2
0
31 Jan 2023
Variants of SGD for Lipschitz Continuous Loss Functions in Low-Precision
  Environments
Variants of SGD for Lipschitz Continuous Loss Functions in Low-Precision Environments
Michael R. Metel
20
1
0
09 Nov 2022
MCTensor: A High-Precision Deep Learning Library with Multi-Component
  Floating-Point
MCTensor: A High-Precision Deep Learning Library with Multi-Component Floating-Point
Tao Yu
Wen-Ping Guo
Jianan Canal Li
Tiancheng Yuan
Chris De Sa
19
4
0
18 Jul 2022
SALO: An Efficient Spatial Accelerator Enabling Hybrid Sparse Attention
  Mechanisms for Long Sequences
SALO: An Efficient Spatial Accelerator Enabling Hybrid Sparse Attention Mechanisms for Long Sequences
Guan Shen
Jieru Zhao
Quan Chen
Jingwen Leng
C. Li
Minyi Guo
39
26
0
29 Jun 2022
Low-Precision Stochastic Gradient Langevin Dynamics
Low-Precision Stochastic Gradient Langevin Dynamics
Ruqi Zhang
A. Wilson
Chris De Sa
BDL
13
14
0
20 Jun 2022
Adaptive Precision Training (AdaPT): A dynamic fixed point quantized
  training approach for DNNs
Adaptive Precision Training (AdaPT): A dynamic fixed point quantized training approach for DNNs
Lorenz Kummer
Kevin Sidak
Tabea Reichmann
Wilfried Gansterer
MQ
19
5
0
28 Jul 2021
Dynamic Neural Network Architectural and Topological Adaptation and
  Related Methods -- A Survey
Dynamic Neural Network Architectural and Topological Adaptation and Related Methods -- A Survey
Lorenz Kummer
AI4CE
32
0
0
28 Jul 2021
How Low Can We Go: Trading Memory for Error in Low-Precision Training
How Low Can We Go: Trading Memory for Error in Low-Precision Training
Chengrun Yang
Ziyang Wu
Jerry Chee
Christopher De Sa
Madeleine Udell
11
2
0
17 Jun 2021
Development of Quantized DNN Library for Exact Hardware Emulation
Development of Quantized DNN Library for Exact Hardware Emulation
M. Kiyama
Motoki Amagasaki
M. Iida
MQ
11
0
0
15 Jun 2021
Hessian Aware Quantization of Spiking Neural Networks
Hessian Aware Quantization of Spiking Neural Networks
H. Lui
Emre Neftci
MQ
8
14
0
29 Apr 2021
Low-Precision Reinforcement Learning: Running Soft Actor-Critic in Half
  Precision
Low-Precision Reinforcement Learning: Running Soft Actor-Critic in Half Precision
Johan Bjorck
Xiangyu Chen
Christopher De Sa
Carla P. Gomes
Kilian Q. Weinberger
13
6
0
26 Feb 2021
PLAM: a Posit Logarithm-Approximate Multiplier
PLAM: a Posit Logarithm-Approximate Multiplier
Raul Murillo
Alberto A. Del Barrio
Guillermo Botella
Min Soo Kim
Hyunjin Kim
N. Bagherzadeh
TPM
24
25
0
18 Feb 2021
Empirical Evaluation of Deep Learning Model Compression Techniques on
  the WaveNet Vocoder
Empirical Evaluation of Deep Learning Model Compression Techniques on the WaveNet Vocoder
Sam Davis
Giuseppe Coccia
Sam Gooch
Julian Mack
4
0
0
20 Nov 2020
Revisiting BFloat16 Training
Revisiting BFloat16 Training
Pedram Zamirai
Jian Zhang
Christopher R. Aberger
Christopher De Sa
FedML
MQ
13
20
0
13 Oct 2020
Low-Complexity LSTM Training and Inference with FloatSD8 Weight
  Representation
Low-Complexity LSTM Training and Inference with FloatSD8 Weight Representation
Yu-Tung Liu
T. Chiueh
MQ
13
1
0
23 Jan 2020
Auto-Precision Scaling for Distributed Deep Learning
Auto-Precision Scaling for Distributed Deep Learning
Ruobing Han
J. Demmel
Yang You
11
5
0
20 Nov 2019
1