Communities
Connect sessions
AI calendar
Organizations
Join Slack
Contact Sales
Search
Open menu
Home
Papers
2503.16672
Cited By
Accelerating Transformer Inference and Training with 2:4 Activation Sparsity
20 March 2025
Daniel Haziza
Timothy Chou
Dhruv Choudhary
Luca Wehrstedt
Francisco Massa
Jiecao Yu
Geonhwa Jeong
Supriya Rao
Patrick Labatut
Jesse Cai
Re-assign community
ArXiv (abs)
PDF
HTML
Papers citing
"Accelerating Transformer Inference and Training with 2:4 Activation Sparsity"
5 / 5 papers shown
Title
Lightweight error mitigation strategies for post-training N:M activation sparsity in LLMs
Shirin Alanova
Kristina Kazistova
Ekaterina Galaeva
Alina Kostromina
Vladimir Smirnov
Redko Dmitry
Alexey Dontsov
Maxim Zhelnin
Evgeny Burnaev
Egor Shvetsov
132
0
0
26 Sep 2025
Amber Pruner: Leveraging N:M Activation Sparsity for Efficient Prefill in Large Language Models
Tai An
Ruwu Cai
Yanzhe Zhang
Yang Liu
Hao Chen
Pengcheng Xie
Sheng Chang
Jing Lin
Gongyi Wang
MoE
124
2
0
04 Aug 2025
TorchAO: PyTorch-Native Training-to-Serving Model Optimization
Andrew Or
Apurva Jain
Daniel Vega-Myhre
Jesse Cai
Charles David Hernandez
...
Christian Puhrsch
Mark Saroufim
Supriya Rao
Thien Tran
Aleksandar Samardžić
MQ
158
4
0
21 Jul 2025
DuoGPT: Training-free Dual Sparsity through Activation-aware Pruning in LLMs
Ruokai Yin
Yuhang Li
Donghyun Lee
Priyadarshini Panda
VLM
190
2
0
25 Jun 2025
SLoPe: Double-Pruned Sparse Plus Lazy Low-Rank Adapter Pretraining of LLMs
International Conference on Learning Representations (ICLR), 2024
Mohammad Mozaffari
Amir Yazdanbakhsh
Zhao Zhang
M. Dehnavi
351
13
0
28 Jan 2025
1