ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2305.12718
  4. Cited By
HighLight: Efficient and Flexible DNN Acceleration with Hierarchical
  Structured Sparsity

HighLight: Efficient and Flexible DNN Acceleration with Hierarchical Structured Sparsity

22 May 2023
Yannan Nellie Wu
Po-An Tsai
Saurav Muralidharan
A. Parashar
Vivienne Sze
J. Emer
ArXivPDFHTML

Papers citing "HighLight: Efficient and Flexible DNN Acceleration with Hierarchical Structured Sparsity"

9 / 9 papers shown
Title
Accelerating LLM Inference with Flexible N:M Sparsity via A Fully Digital Compute-in-Memory Accelerator
Accelerating LLM Inference with Flexible N:M Sparsity via A Fully Digital Compute-in-Memory Accelerator
Akshat Ramachandran
Souvik Kundu
Arnab Raha
Shamik Kundu
Deepak K. Mathaikutty
Tushar Krishna
27
1
0
19 Apr 2025
BitMoD: Bit-serial Mixture-of-Datatype LLM Acceleration
Yuzong Chen
Ahmed F. AbouElhamayed
Xilai Dai
Yang Wang
Marta Andronic
G. Constantinides
Mohamed S. Abdelfattah
MQ
103
1
0
18 Nov 2024
Towards Efficient Neuro-Symbolic AI: From Workload Characterization to
  Hardware Architecture
Towards Efficient Neuro-Symbolic AI: From Workload Characterization to Hardware Architecture
Zishen Wan
Che-Kai Liu
Hanchen Yang
Ritik Raj
Chaojian Li
...
Yingyan Celine Lin
Mohamed Ibrahim
Jan M. Rabaey
Tushar Krishna
A. Raychowdhury
39
8
0
20 Sep 2024
Toward Efficient Permutation for Hierarchical N:M Sparsity on GPUs
Toward Efficient Permutation for Hierarchical N:M Sparsity on GPUs
Seungmin Yu
Xiaodie Yi
Hayun Lee
Dongkun Shin
24
1
0
30 Jul 2024
SDQ: Sparse Decomposed Quantization for LLM Inference
SDQ: Sparse Decomposed Quantization for LLM Inference
Geonhwa Jeong
Po-An Tsai
S. Keckler
Tushar Krishna
MQ
35
3
0
19 Jun 2024
Torch2Chip: An End-to-end Customizable Deep Neural Network Compression
  and Deployment Toolkit for Prototype Hardware Accelerator Design
Torch2Chip: An End-to-end Customizable Deep Neural Network Compression and Deployment Toolkit for Prototype Hardware Accelerator Design
Jian Meng
Yuan Liao
Anupreetham Anupreetham
Ahmed Hassan
Shixing Yu
Han-Sok Suh
Xiaofeng Hu
Jae-sun Seo
MQ
49
1
0
02 May 2024
Abstracting Sparse DNN Acceleration via Structured Sparse Tensor
  Decomposition
Abstracting Sparse DNN Acceleration via Structured Sparse Tensor Decomposition
Geonhwa Jeong
Po-An Tsai
A. Bambhaniya
S. Keckler
Tushar Krishna
25
7
0
12 Mar 2024
Sparsity in Deep Learning: Pruning and growth for efficient inference
  and training in neural networks
Sparsity in Deep Learning: Pruning and growth for efficient inference and training in neural networks
Torsten Hoefler
Dan Alistarh
Tal Ben-Nun
Nikoli Dryden
Alexandra Peste
MQ
141
684
0
31 Jan 2021
What is the State of Neural Network Pruning?
What is the State of Neural Network Pruning?
Davis W. Blalock
Jose Javier Gonzalez Ortiz
Jonathan Frankle
John Guttag
185
1,027
0
06 Mar 2020
1