ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2026 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1709.04060
  4. Cited By
Streamlined Deployment for Quantized Neural Networks
v1v2 (latest)

Streamlined Deployment for Quantized Neural Networks

12 September 2017
Yaman Umuroglu
Magnus Jahre
    MQ
ArXiv (abs)PDFHTML

Papers citing "Streamlined Deployment for Quantized Neural Networks"

12 / 12 papers shown
Real-Time Multi-Object Tracking using YOLOv8 and SORT on a SoC FPGA
Real-Time Multi-Object Tracking using YOLOv8 and SORT on a SoC FPGAInternational Workshop on Applied Reconfigurable Computing (ARC), 2025
Michal Danilowicz
T. Kryjak
VOT
307
5
0
17 Mar 2025
Fast, Scalable, Energy-Efficient Non-element-wise Matrix Multiplication
  on FPGA
Fast, Scalable, Energy-Efficient Non-element-wise Matrix Multiplication on FPGA
Xuqi Zhu
Huaizhi Zhang
JunKyu Lee
Jiacheng Zhu
Chandrajit Pal
S. Saha
Klaus D. McDonald-Maier
X. Zhai
254
0
0
02 Jul 2024
A2Q+: Improving Accumulator-Aware Weight Quantization
A2Q+: Improving Accumulator-Aware Weight Quantization
Ian Colbert
Alessandro Pappalardo
Jakoba Petri-Koenig
Yaman Umuroglu
MQ
262
11
0
19 Jan 2024
A2Q: Accumulator-Aware Quantization with Guaranteed Overflow Avoidance
A2Q: Accumulator-Aware Quantization with Guaranteed Overflow AvoidanceIEEE International Conference on Computer Vision (ICCV), 2023
Ian Colbert
Alessandro Pappalardo
Jakoba Petri-Koenig
MQ
311
19
0
25 Aug 2023
Open-source FPGA-ML codesign for the MLPerf Tiny Benchmark
Open-source FPGA-ML codesign for the MLPerf Tiny Benchmark
H. Borras
G. D. Guglielmo
Javier Mauricio Duarte
Nicolò Ghielmetti
B. Hawks
...
Nhan Tran
Yaman Umuroglu
Olivia Weng
Aidan Yokuda
Michaela Blott
VLMMQ
199
18
0
23 Jun 2022
Applications and Techniques for Fast Machine Learning in Science
Applications and Techniques for Fast Machine Learning in ScienceFrontiers in Big Data (Front. Big Data), 2021
A. Deiana
Nhan Tran
Joshua C. Agar
Michaela Blott
G. D. Guglielmo
...
Ashish Sharma
S. Summers
Pietro Vischia
J. Vlimant
Olivia Weng
294
86
0
25 Oct 2021
Benchmarking Quantized Neural Networks on FPGAs with FINN
Benchmarking Quantized Neural Networks on FPGAs with FINN
Quentin Ducasse
Pascal Cotret
Loïc Lagadec
Rob Stewart
MQ
158
16
0
02 Feb 2021
Diagnostic data integration using deep neural networks for real-time
  plasma analysis
Diagnostic data integration using deep neural networks for real-time plasma analysisIEEE Transactions on Nuclear Science (TNS), 2020
A. R. Garola
R. Cavazzana
M. Gobbin
R. Delogu
G. Manduchi
C. Taliercio
A. Luchetta
235
6
0
28 Oct 2020
Resource-Efficient Speech Mask Estimation for Multi-Channel Speech
  Enhancement
Resource-Efficient Speech Mask Estimation for Multi-Channel Speech Enhancement
Lukas Pfeifenberger
Matthias Zöhrer
Günther Schindler
Wolfgang Roth
Holger Fröning
Franz Pernkopf
142
1
0
22 Jul 2020
Quantized Neural Network Inference with Precision Batching
Quantized Neural Network Inference with Precision Batching
Maximilian Lam
Zachary Yedidia
Colby R. Banbury
Vijay Janapa Reddi
MQ
203
1
0
26 Feb 2020
Memory-Driven Mixed Low Precision Quantization For Enabling Deep Network
  Inference On Microcontrollers
Memory-Driven Mixed Low Precision Quantization For Enabling Deep Network Inference On MicrocontrollersConference on Machine Learning and Systems (MLSys), 2019
Manuele Rusci
Alessandro Capotondi
Luca Benini
MQ
360
86
0
30 May 2019
High performance ultra-low-precision convolutions on mobile devices
High performance ultra-low-precision convolutions on mobile devices
Andrew Tulloch
Yangqing Jia
HAIMQ
209
29
0
06 Dec 2017
1
Page 1 of 1