Communities
Connect sessions
AI calendar
Organizations
Join Slack
Contact Sales
Search
Open menu
Home
Papers
All Papers
0 / 0 papers shown
Title
Home
Papers
1812.08011
Cited By
Training Deep Neural Networks with 8-bit Floating Point Numbers
19 December 2018
Naigang Wang
Jungwook Choi
D. Brand
Chia-Yu Chen
K. Gopalakrishnan
MQ
Re-assign community
ArXiv (abs)
PDF
HTML
Papers citing
"Training Deep Neural Networks with 8-bit Floating Point Numbers"
50 / 212 papers shown
Title
Productivity meets Performance: Julia on A64FX
IEEE International Conference on Cluster Computing (Cluster), 2022
Mosè Giordano
Milan Klower
Valentin Churavy
78
11
0
26 Jul 2022
CTMQ: Cyclic Training of Convolutional Neural Networks with Multiple Quantization Steps
Hyunjin Kim
Jungwoon Shin
Alberto A. Del Barrio
MQ
131
2
0
26 Jun 2022
GACT: Activation Compressed Training for Generic Network Architectures
International Conference on Machine Learning (ICML), 2022
Xiaoxuan Liu
Lianmin Zheng
Yi Xu
Yukuo Cen
Weize Chen
...
Zhiyuan Liu
Jie Tang
Joey Gonzalez
Michael W. Mahoney
Alvin Cheung
VLM
GNN
MQ
209
38
0
22 Jun 2022
Low-Precision Stochastic Gradient Langevin Dynamics
International Conference on Machine Learning (ICML), 2022
Ruqi Zhang
A. Wilson
Chris De Sa
BDL
117
18
0
20 Jun 2022
Optimal Clipping and Magnitude-aware Differentiation for Improved Quantization-aware Training
International Conference on Machine Learning (ICML), 2022
Charbel Sakr
Steve Dai
Rangharajan Venkatesan
B. Zimmer
W. Dally
Brucek Khailany
MQ
135
49
0
13 Jun 2022
8-bit Numerical Formats for Deep Neural Networks
Badreddine Noune
Philip Jones
Daniel Justus
Dominic Masters
Carlo Luschi
MQ
142
37
0
06 Jun 2022
BiT: Robustly Binarized Multi-distilled Transformer
Neural Information Processing Systems (NeurIPS), 2022
Zechun Liu
Barlas Oğuz
Aasish Pappu
Lin Xiao
Scott Yih
Meng Li
Raghuraman Krishnamoorthi
Yashar Mehdad
MQ
212
73
0
25 May 2022
Structural Dropout for Model Width Compression
Julian Knodt
OffRL
70
1
0
13 May 2022
Neural Architecture Search using Property Guided Synthesis
Charles Jin
P. Phothilimthana
Sudip Roy
123
7
0
08 May 2022
Schrödinger's FP: Dynamic Adaptation of Floating-Point Containers for Deep Learning Training
Milovs Nikolić
Enrique Torres Sanchez
Jia-Hui Wang
Ali Hadi Zadeh
Mostafa Mahmoud
Ameer Abdelhadi
Kareem Ibrahim
Andreas Moshovos
MQ
126
1
0
28 Apr 2022
Vision Transformer Compression with Structured Pruning and Low Rank Approximation
Ankur Kumar
ViT
55
6
0
25 Mar 2022
Hardware Approximate Techniques for Deep Neural Network Accelerators: A Survey
ACM Computing Surveys (ACM CSUR), 2022
Giorgos Armeniakos
Georgios Zervakis
Dimitrios Soudris
J. Henkel
441
116
0
16 Mar 2022
FlexBlock: A Flexible DNN Training Accelerator with Multi-Mode Block Floating Point Support
IEEE transactions on computers (IEEE Trans. Comput.), 2022
Seock-Hwan Noh
Jahyun Koo
Seunghyun Lee
Jongse Park
Jaeha Kung
AI4CE
132
25
0
13 Mar 2022
On the influence of stochastic roundoff errors and their bias on the convergence of the gradient descent method with low-precision floating-point computation
Lu Xia
Stefano Massei
M. Hochstenbach
B. Koren
88
5
0
24 Feb 2022
Resource-Efficient Deep Learning: A Survey on Model-, Arithmetic-, and Implementation-Level Techniques
ACM Computing Surveys (CSUR), 2021
JunKyu Lee
L. Mukhanov
A. S. Molahosseini
U. Minhas
Yang Hua
Jesus Martinez del Rincon
K. Dichev
Cheol-Ho Hong
Hans Vandierendonck
129
35
0
30 Dec 2021
Accurate Neural Training with 4-bit Matrix Multiplications at Standard Formats
International Conference on Learning Representations (ICLR), 2021
Brian Chmiel
Ron Banner
Elad Hoffer
Hilla Ben Yaacov
Daniel Soudry
MQ
249
28
0
19 Dec 2021
LVAC: Learned Volumetric Attribute Compression for Point Clouds using Coordinate Based Networks
Berivan Isik
P. Chou
S. Hwang
Nick Johnston
G. Toderici
3DPC
182
31
0
17 Nov 2021
DAdaQuant: Doubly-adaptive quantization for communication-efficient Federated Learning
International Conference on Machine Learning (ICML), 2021
Robert Hönig
Yiren Zhao
Robert D. Mullins
FedML
289
67
0
31 Oct 2021
NeRV: Neural Representations for Videos
Hao Chen
Bo He
Hanyu Wang
Yixuan Ren
Ser-Nam Lim
Abhinav Shrivastava
123
305
0
26 Oct 2021
Exploring System Performance of Continual Learning for Mobile and Embedded Sensing Applications
Young D. Kwon
Jagmohan Chauhan
Abhishek Kumar
Pan Hui
Cecilia Mascolo
CLL
HAI
148
34
0
25 Oct 2021
LightSeq2: Accelerated Training for Transformer-based Models on GPUs
International Conference for High Performance Computing, Networking, Storage and Analysis (SC), 2021
Xiaohui Wang
Yang Wei
Ying Xiong
Guyue Huang
Xian Qian
Yufei Ding
Mingxuan Wang
Lei Li
VLM
163
36
0
12 Oct 2021
Shift-BNN: Highly-Efficient Probabilistic Bayesian Neural Network Training via Memory-Friendly Pattern Retrieving
Qiyu Wan
Haojun Xia
Xingyao Zhang
Lening Wang
Shuaiwen Leon Song
Xin Fu
OOD
98
9
0
07 Oct 2021
8-bit Optimizers via Block-wise Quantization
Tim Dettmers
M. Lewis
Sam Shleifer
Luke Zettlemoyer
MQ
304
360
0
06 Oct 2021
SDR: Efficient Neural Re-ranking using Succinct Document Representation
Nachshon Cohen
Amit Portnoy
B. Fetahu
A. Ingber
AI4TS
182
11
0
03 Oct 2021
Artificial Intelligence in the Low-Level Realm -- A Survey
Vahid Mohammadi Safarzadeh
Hamed Ghasr Loghmani
88
1
0
19 Sep 2021
Efficient Visual Recognition with Deep Neural Networks: A Survey on Recent Advances and New Directions
Machine Intelligence Research (MIR), 2021
Yang Wu
Dingheng Wang
Xiaotong Lu
Fan Yang
Guoqi Li
Weiming Dong
Jianbo Shi
237
18
0
30 Aug 2021
Estimating Counts Through an Average Rounded to the Nearest Non-negative Integer and its Theoretical & Practical Effects
R. Rivera
Axel Cortes-Cubero
Roberto Reyes-Carranza
W. Rolke
103
0
0
04 Jul 2021
Secure Quantized Training for Deep Learning
Marcel Keller
Ke Sun
MQ
128
72
0
01 Jul 2021
Reducing numerical precision preserves classification accuracy in Mondrian Forests
Marc Vicuna
Martin Khannouz
Gregory Kiar
Yohan Chatelain
Tristan Glatard
MQ
96
3
0
28 Jun 2021
LNS-Madam: Low-Precision Training in Logarithmic Number System using Multiplicative Weight Update
IEEE transactions on computers (IEEE Trans. Comput.), 2021
Jiawei Zhao
Steve Dai
Rangharajan Venkatesan
Brian Zimmer
Mustafa Ali
Xuan Li
Brucek Khailany
B. Dally
Anima Anandkumar
MQ
111
18
0
26 Jun 2021
Towards Efficient Full 8-bit Integer DNN Online Training on Resource-limited Devices without Batch Normalization
Yukuan Yang
Xiaowei Chi
Lei Deng
Tianyi Yan
Feng Gao
Guoqi Li
MQ
143
6
0
27 May 2021
ActNN: Reducing Training Memory Footprint via 2-Bit Activation Compressed Training
International Conference on Machine Learning (ICML), 2021
Jianfei Chen
Lianmin Zheng
Z. Yao
Yi Xu
Ion Stoica
Michael W. Mahoney
Joseph E. Gonzalez
MQ
153
86
0
29 Apr 2021
ScaleCom: Scalable Sparsified Gradient Compression for Communication-Efficient Distributed Training
Neural Information Processing Systems (NeurIPS), 2021
Chia-Yu Chen
Jiamin Ni
Songtao Lu
Xiaodong Cui
Pin-Yu Chen
...
Naigang Wang
Swagath Venkataramani
Vijayalakshmi Srinivasan
Wei Zhang
K. Gopalakrishnan
155
71
0
21 Apr 2021
Optimal Size-Performance Tradeoffs: Weighing PoS Tagger Models
Magnus Jacobsen
Mikkel H. Sorensen
Leon Derczynski
126
4
0
16 Apr 2021
All-You-Can-Fit 8-Bit Flexible Floating-Point Format for Accurate and Memory-Efficient Inference of Deep Neural Networks
Cheng-Wei Huang
Tim-Wei Chen
Juinn-Dar Huang
MQ
88
6
0
15 Apr 2021
Distributed Learning Systems with First-order Methods
Ji Liu
Ce Zhang
101
46
0
12 Apr 2021
Charged particle tracking via edge-classifying interaction networks
Computing and Software for Big Science (CSBS), 2021
G. Dezoort
S. Thais
Javier Mauricio Duarte
Vesal Razavimaleki
M. Atkinson
I. Ojalvo
Mark S. Neubauer
P. Elmer
169
52
0
30 Mar 2021
A Simple and Efficient Stochastic Rounding Method for Training Neural Networks in Low Precision
Lu Xia
M. Anthonissen
M. Hochstenbach
B. Koren
BDL
38
9
0
24 Mar 2021
Low-Precision Reinforcement Learning: Running Soft Actor-Critic in Half Precision
International Conference on Machine Learning (ICML), 2021
Johan Bjorck
Xiangyu Chen
Christopher De Sa
Daniel Schwalbe-Koda
Kilian Q. Weinberger
194
5
0
26 Feb 2021
Ps and Qs: Quantization-aware pruning for efficient low latency neural network inference
Frontiers in Artificial Intelligence (Front. Artif. Intell.), 2021
B. Hawks
Javier Mauricio Duarte
Nicholas J. Fraser
Alessandro Pappalardo
N. Tran
Yaman Umuroglu
MQ
167
62
0
22 Feb 2021
Training Neural Networks is
∃
R
\exists\mathbb R
∃
R
-complete
Mikkel Abrahamsen
Linda Kleist
Tillmann Miltzow
106
1
0
19 Feb 2021
NEAT: A Framework for Automated Exploration of Floating Point Approximations
Saeid Barati
Lee Ehudin
Hank Hoffmann
60
1
0
17 Feb 2021
GradPIM: A Practical Processing-in-DRAM Architecture for Gradient Descent
International Symposium on High-Performance Computer Architecture (HPCA), 2021
Heesu Kim
Hanmin Park
Taehyun Kim
Kwanheum Cho
Eojin Lee
Soojung Ryu
Hyuk-Jae Lee
Kiyoung Choi
Jinho Lee
122
47
0
15 Feb 2021
Distribution Adaptive INT8 Quantization for Training CNNs
AAAI Conference on Artificial Intelligence (AAAI), 2021
Kang Zhao
Sida Huang
Pan Pan
Yinghan Li
Yingya Zhang
Zhenyu Gu
Yinghui Xu
MQ
155
76
0
09 Feb 2021
Enabling Binary Neural Network Training on the Edge
Erwei Wang
James J. Davis
Daniele Moro
Piotr Zielinski
Jia Jie Lim
C. Coelho
S. Chatterjee
P. Cheung
George A. Constantinides
MQ
433
30
0
08 Feb 2021
EFloat: Entropy-coded Floating Point Format for Compressing Vector Embedding Models
R. Bordawekar
B. Abali
Ming-Hung Chen
MQ
107
3
0
04 Feb 2021
Pruning and Quantization for Deep Neural Network Acceleration: A Survey
Neurocomputing (Neurocomputing), 2021
Tailin Liang
C. Glossner
Lei Wang
Shaobo Shi
Xiaotong Zhang
MQ
396
817
0
24 Jan 2021
Old but Gold: Reconsidering the value of feedforward learners for software analytics
Rahul Yedida
Xueqi Yang
Tim Menzies
AI4TS
96
4
0
15 Jan 2021
SmartDeal: Re-Modeling Deep Network Weights for Efficient Inference and Training
IEEE Transactions on Neural Networks and Learning Systems (TNNLS), 2021
Xiaohan Chen
Yang Zhao
Yue Wang
Pengfei Xu
Haoran You
Chaojian Li
Y. Fu
Yingyan Lin
Zinan Lin
251
1
0
04 Jan 2021
FracTrain: Fractionally Squeezing Bit Savings Both Temporally and Spatially for Efficient DNN Training
Neural Information Processing Systems (NeurIPS), 2020
Y. Fu
Haoran You
Yang Zhao
Yue Wang
Chaojian Li
K. Gopalakrishnan
Zinan Lin
Yingyan Lin
MQ
213
34
0
24 Dec 2020
Previous
1
2
3
4
5
Next