ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1712.01507
  4. Cited By
Bit Fusion: Bit-Level Dynamically Composable Architecture for
  Accelerating Deep Neural Networks

Bit Fusion: Bit-Level Dynamically Composable Architecture for Accelerating Deep Neural Networks

5 December 2017
Hardik Sharma
Jongse Park
Naveen Suda
Liangzhen Lai
Benson Chau
J. Kim
Vikas Chandra
H. Esmaeilzadeh
    MQ
ArXivPDFHTML

Papers citing "Bit Fusion: Bit-Level Dynamically Composable Architecture for Accelerating Deep Neural Networks"

44 / 44 papers shown
Title
LightNobel: Improving Sequence Length Limitation in Protein Structure Prediction Model via Adaptive Activation Quantization
LightNobel: Improving Sequence Length Limitation in Protein Structure Prediction Model via Adaptive Activation Quantization
Seunghee Han
S. Choi
J. Kim
26
0
0
09 May 2025
HALO: Hardware-aware quantization with low critical-path-delay weights for LLM acceleration
HALO: Hardware-aware quantization with low critical-path-delay weights for LLM acceleration
Rohan Juneja
Shivam Aggarwal
Safeen Huda
Tulika Mitra
L. Peh
45
0
0
27 Feb 2025
Ditto: Accelerating Diffusion Model via Temporal Value Similarity
Ditto: Accelerating Diffusion Model via Temporal Value Similarity
Sungbin Kim
Hyunwuk Lee
Wonho Cho
Mincheol Park
Won Woo Ro
58
1
0
20 Jan 2025
BF-IMNA: A Bit Fluid In-Memory Neural Architecture for Neural Network
  Acceleration
BF-IMNA: A Bit Fluid In-Memory Neural Architecture for Neural Network Acceleration
M. Rakka
Rachid Karami
A. Eltawil
M. Fouda
Fadi J. Kurdahi
MQ
39
1
0
03 Nov 2024
ARQ: A Mixed-Precision Quantization Framework for Accurate and Certifiably Robust DNNs
ARQ: A Mixed-Precision Quantization Framework for Accurate and Certifiably Robust DNNs
Yuchen Yang
Shubham Ugare
Yifan Zhao
Gagandeep Singh
Sasa Misailovic
MQ
26
0
0
31 Oct 2024
Dishonest Approximate Computing: A Coming Crisis for Cloud Clients
Dishonest Approximate Computing: A Coming Crisis for Cloud Clients
Ye Wang
Jian Dong
Ming Han
Jin Wu
Gang Qu
22
0
0
24 May 2024
Exploring Quantization and Mapping Synergy in Hardware-Aware Deep Neural Network Accelerators
Exploring Quantization and Mapping Synergy in Hardware-Aware Deep Neural Network Accelerators
Jan Klhufek
Miroslav Safar
Vojtěch Mrázek
Z. Vašíček
Lukás Sekanina
MQ
32
1
0
08 Apr 2024
DyBit: Dynamic Bit-Precision Numbers for Efficient Quantized Neural
  Network Inference
DyBit: Dynamic Bit-Precision Numbers for Efficient Quantized Neural Network Inference
Jiajun Zhou
Jiajun Wu
Yizhao Gao
Yuhao Ding
Chaofan Tao
Bo-wen Li
Fengbin Tu
Kwang-Ting Cheng
Hayden Kwok-Hay So
Ngai Wong
MQ
18
7
0
24 Feb 2023
Fixflow: A Framework to Evaluate Fixed-point Arithmetic in Light-Weight
  CNN Inference
Fixflow: A Framework to Evaluate Fixed-point Arithmetic in Light-Weight CNN Inference
Farhad Taheri
Siavash Bayat Sarmadi
H. Mosanaei-Boorani
Reza Taheri
MQ
18
1
0
19 Feb 2023
The Hidden Power of Pure 16-bit Floating-Point Neural Networks
The Hidden Power of Pure 16-bit Floating-Point Neural Networks
Juyoung Yun
Byungkon Kang
Zhoulai Fu
MQ
23
1
0
30 Jan 2023
Does Federated Learning Really Need Backpropagation?
Does Federated Learning Really Need Backpropagation?
H. Feng
Tianyu Pang
Chao Du
Wei-Neng Chen
Shuicheng Yan
Min-Bin Lin
FedML
26
10
0
28 Jan 2023
ANT: Exploiting Adaptive Numerical Data Type for Low-bit Deep Neural
  Network Quantization
ANT: Exploiting Adaptive Numerical Data Type for Low-bit Deep Neural Network Quantization
Cong Guo
Chen Zhang
Jingwen Leng
Zihan Liu
Fan Yang
Yun-Bo Liu
Minyi Guo
Yuhao Zhu
MQ
16
55
0
30 Aug 2022
Mixed-Precision Neural Networks: A Survey
Mixed-Precision Neural Networks: A Survey
M. Rakka
M. Fouda
Pramod P. Khargonekar
Fadi J. Kurdahi
MQ
21
11
0
11 Aug 2022
Design of High-Throughput Mixed-Precision CNN Accelerators on FPGA
Design of High-Throughput Mixed-Precision CNN Accelerators on FPGA
Cecilia Latotzke
Tim Ciesielski
T. Gemmeke
MQ
13
7
0
09 Aug 2022
Symmetry Regularization and Saturating Nonlinearity for Robust
  Quantization
Symmetry Regularization and Saturating Nonlinearity for Robust Quantization
Sein Park
Yeongsang Jang
Eunhyeok Park
MQ
14
1
0
31 Jul 2022
Accelerating Attention through Gradient-Based Learned Runtime Pruning
Accelerating Attention through Gradient-Based Learned Runtime Pruning
Zheng Li
Soroush Ghodrati
Amir Yazdanbakhsh
H. Esmaeilzadeh
Mingu Kang
19
17
0
07 Apr 2022
FlexBlock: A Flexible DNN Training Accelerator with Multi-Mode Block
  Floating Point Support
FlexBlock: A Flexible DNN Training Accelerator with Multi-Mode Block Floating Point Support
Seock-Hwan Noh
Jahyun Koo
Seunghyun Lee
Jongse Park
Jaeha Kung
AI4CE
29
17
0
13 Mar 2022
EcoFlow: Efficient Convolutional Dataflows for Low-Power Neural Network
  Accelerators
EcoFlow: Efficient Convolutional Dataflows for Low-Power Neural Network Accelerators
Lois Orosa
Skanda Koppula
Yaman Umuroglu
Konstantinos Kanellopoulos
Juan Gómez Luna
Michaela Blott
K. Vissers
O. Mutlu
38
4
0
04 Feb 2022
SPA-GCN: Efficient and Flexible GCN Accelerator with an Application for
  Graph Similarity Computation
SPA-GCN: Efficient and Flexible GCN Accelerator with an Application for Graph Similarity Computation
Atefeh Sohrabizadeh
Yuze Chi
Jason Cong
GNN
29
1
0
10 Nov 2021
Transformer Acceleration with Dynamic Sparse Attention
Transformer Acceleration with Dynamic Sparse Attention
Liu Liu
Zheng Qu
Zhaodong Chen
Yufei Ding
Yuan Xie
19
20
0
21 Oct 2021
Towards Mixed-Precision Quantization of Neural Networks via Constrained
  Optimization
Towards Mixed-Precision Quantization of Neural Networks via Constrained Optimization
Weihan Chen
Peisong Wang
Jian Cheng
MQ
42
61
0
13 Oct 2021
2-in-1 Accelerator: Enabling Random Precision Switch for Winning Both Adversarial Robustness and Efficiency
2-in-1 Accelerator: Enabling Random Precision Switch for Winning Both Adversarial Robustness and Efficiency
Yonggan Fu
Yang Katie Zhao
Qixuan Yu
Chaojian Li
Yingyan Lin
AAML
44
12
0
11 Sep 2021
MOHAQ: Multi-Objective Hardware-Aware Quantization of Recurrent Neural
  Networks
MOHAQ: Multi-Objective Hardware-Aware Quantization of Recurrent Neural Networks
Nesma M. Rezk
Tomas Nordstrom
D. Stathis
Z. Ul-Abdin
E. Aksoy
A. Hemani
MQ
20
1
0
02 Aug 2021
On the Impact of Device-Level Techniques on Energy-Efficiency of Neural
  Network Accelerators
On the Impact of Device-Level Techniques on Energy-Efficiency of Neural Network Accelerators
Seyed Morteza Nabavinejad
Behzad Salami
15
1
0
26 Jun 2021
Random and Adversarial Bit Error Robustness: Energy-Efficient and Secure
  DNN Accelerators
Random and Adversarial Bit Error Robustness: Energy-Efficient and Secure DNN Accelerators
David Stutz
Nandhini Chandramoorthy
Matthias Hein
Bernt Schiele
AAML
MQ
22
18
0
16 Apr 2021
Enabling Design Methodologies and Future Trends for Edge AI:
  Specialization and Co-design
Enabling Design Methodologies and Future Trends for Edge AI: Specialization and Co-design
Cong Hao
Jordan Dotzel
Jinjun Xiong
Luca Benini
Zhiru Zhang
Deming Chen
50
34
0
25 Mar 2021
FracTrain: Fractionally Squeezing Bit Savings Both Temporally and Spatially for Efficient DNN Training
FracTrain: Fractionally Squeezing Bit Savings Both Temporally and Spatially for Efficient DNN Training
Y. Fu
Haoran You
Yang Katie Zhao
Yue Wang
Chaojian Li
K. Gopalakrishnan
Zhangyang Wang
Yingyan Lin
MQ
30
32
0
24 Dec 2020
Hardware and Software Optimizations for Accelerating Deep Neural
  Networks: Survey of Current Trends, Challenges, and the Road Ahead
Hardware and Software Optimizations for Accelerating Deep Neural Networks: Survey of Current Trends, Challenges, and the Road Ahead
Maurizio Capra
Beatrice Bussolino
Alberto Marchisio
Guido Masera
Maurizio Martina
Muhammad Shafique
BDL
56
140
0
21 Dec 2020
FantastIC4: A Hardware-Software Co-Design Approach for Efficiently
  Running 4bit-Compact Multilayer Perceptrons
FantastIC4: A Hardware-Software Co-Design Approach for Efficiently Running 4bit-Compact Multilayer Perceptrons
Simon Wiedemann
Suhas Shivapakash
P. Wiedemann
Daniel Becking
Wojciech Samek
F. Gerfers
Thomas Wiegand
MQ
18
7
0
17 Dec 2020
SpAtten: Efficient Sparse Attention Architecture with Cascade Token and
  Head Pruning
SpAtten: Efficient Sparse Attention Architecture with Cascade Token and Head Pruning
Hanrui Wang
Zhekai Zhang
Song Han
26
374
0
17 Dec 2020
Mix and Match: A Novel FPGA-Centric Deep Neural Network Quantization
  Framework
Mix and Match: A Novel FPGA-Centric Deep Neural Network Quantization Framework
Sung-En Chang
Yanyu Li
Mengshu Sun
Runbin Shi
Hayden Kwok-Hay So
Xuehai Qian
Yanzhi Wang
Xue Lin
MQ
18
82
0
08 Dec 2020
Tensor Casting: Co-Designing Algorithm-Architecture for Personalized
  Recommendation Training
Tensor Casting: Co-Designing Algorithm-Architecture for Personalized Recommendation Training
Youngeun Kwon
Yunjae Lee
Minsoo Rhu
19
39
0
25 Oct 2020
FPRaker: A Processing Element For Accelerating Neural Network Training
FPRaker: A Processing Element For Accelerating Neural Network Training
Omar Mohamed Awad
Mostafa Mahmoud
Isak Edo Vivancos
Ali Hadi Zadeh
Ciaran Bannon
Anand Jayarajan
Gennady Pekhimenko
Andreas Moshovos
20
15
0
15 Oct 2020
Hardware Acceleration of Sparse and Irregular Tensor Computations of ML
  Models: A Survey and Insights
Hardware Acceleration of Sparse and Irregular Tensor Computations of ML Models: A Survey and Insights
Shail Dave
Riyadh Baghdadi
Tony Nowatzki
Sasikanth Avancha
Aviral Shrivastava
Baoxin Li
46
81
0
02 Jul 2020
APQ: Joint Search for Network Architecture, Pruning and Quantization
  Policy
APQ: Joint Search for Network Architecture, Pruning and Quantization Policy
Tianzhe Wang
Kuan-Chieh Jackson Wang
Han Cai
Ji Lin
Zhijian Liu
Song Han
MQ
36
174
0
15 Jun 2020
EDD: Efficient Differentiable DNN Architecture and Implementation
  Co-search for Embedded AI Solutions
EDD: Efficient Differentiable DNN Architecture and Implementation Co-search for Embedded AI Solutions
Yuhong Li
Cong Hao
Xiaofan Zhang
Xinheng Liu
Yao Chen
Jinjun Xiong
Wen-mei W. Hwu
Deming Chen
32
77
0
06 May 2020
Efficient Bitwidth Search for Practical Mixed Precision Neural Network
Efficient Bitwidth Search for Practical Mixed Precision Neural Network
Yuhang Li
Wei Wang
Haoli Bai
Ruihao Gong
Xin Dong
F. Yu
MQ
13
20
0
17 Mar 2020
TMA: Tera-MACs/W Neural Hardware Inference Accelerator with a
  Multiplier-less Massive Parallel Processor
TMA: Tera-MACs/W Neural Hardware Inference Accelerator with a Multiplier-less Massive Parallel Processor
Hyunbin Park
Dohyun Kim
Shiho Kim
BDL
19
1
0
08 Sep 2019
NeuPart: Using Analytical Models to Drive Energy-Efficient Partitioning
  of CNN Computations on Cloud-Connected Mobile Clients
NeuPart: Using Analytical Models to Drive Energy-Efficient Partitioning of CNN Computations on Cloud-Connected Mobile Clients
Susmita Dey Manasi
F. S. Snigdha
S. Sapatnekar
26
16
0
09 May 2019
AutoQ: Automated Kernel-Wise Neural Network Quantization
AutoQ: Automated Kernel-Wise Neural Network Quantization
Qian Lou
Feng Guo
Lantao Liu
Minje Kim
Lei Jiang
MQ
19
97
0
15 Feb 2019
HAQ: Hardware-Aware Automated Quantization with Mixed Precision
HAQ: Hardware-Aware Automated Quantization with Mixed Precision
Kuan-Chieh Jackson Wang
Zhijian Liu
Yujun Lin
Ji Lin
Song Han
MQ
41
872
0
21 Nov 2018
Morph: Flexible Acceleration for 3D CNN-based Video Understanding
Morph: Flexible Acceleration for 3D CNN-based Video Understanding
Kartik Hegde
R. Agrawal
Yulun Yao
Christopher W. Fletcher
25
70
0
16 Oct 2018
Towards Fast and Energy-Efficient Binarized Neural Network Inference on
  FPGA
Towards Fast and Energy-Efficient Binarized Neural Network Inference on FPGA
Cheng Fu
Shilin Zhu
Hao Su
Ching-En Lee
Jishen Zhao
MQ
15
31
0
04 Oct 2018
GANAX: A Unified MIMD-SIMD Acceleration for Generative Adversarial
  Networks
GANAX: A Unified MIMD-SIMD Acceleration for Generative Adversarial Networks
Amir Yazdanbakhsh
Hajar Falahati
Philip J. Wolfe
K. Samadi
N. Kim
H. Esmaeilzadeh
20
71
0
10 May 2018
1