ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2503.16672
  4. Cited By
Accelerating Transformer Inference and Training with 2:4 Activation Sparsity

Accelerating Transformer Inference and Training with 2:4 Activation Sparsity

20 March 2025
Daniel Haziza
Timothy Chou
Dhruv Choudhary
Luca Wehrstedt
Francisco Massa
Jiecao Yu
Geonhwa Jeong
Supriya Rao
Patrick Labatut
Jesse Cai
ArXiv (abs)PDFHTML

Papers citing "Accelerating Transformer Inference and Training with 2:4 Activation Sparsity"

5 / 5 papers shown
Title
Lightweight error mitigation strategies for post-training N:M activation sparsity in LLMs
Lightweight error mitigation strategies for post-training N:M activation sparsity in LLMs
Shirin Alanova
Kristina Kazistova
Ekaterina Galaeva
Alina Kostromina
Vladimir Smirnov
Redko Dmitry
Alexey Dontsov
Maxim Zhelnin
Evgeny Burnaev
Egor Shvetsov
132
0
0
26 Sep 2025
Amber Pruner: Leveraging N:M Activation Sparsity for Efficient Prefill in Large Language Models
Amber Pruner: Leveraging N:M Activation Sparsity for Efficient Prefill in Large Language Models
Tai An
Ruwu Cai
Yanzhe Zhang
Yang Liu
Hao Chen
Pengcheng Xie
Sheng Chang
Jing Lin
Gongyi Wang
MoE
124
2
0
04 Aug 2025
TorchAO: PyTorch-Native Training-to-Serving Model Optimization
TorchAO: PyTorch-Native Training-to-Serving Model Optimization
Andrew Or
Apurva Jain
Daniel Vega-Myhre
Jesse Cai
Charles David Hernandez
...
Christian Puhrsch
Mark Saroufim
Supriya Rao
Thien Tran
Aleksandar Samardžić
MQ
158
4
0
21 Jul 2025
DuoGPT: Training-free Dual Sparsity through Activation-aware Pruning in LLMs
DuoGPT: Training-free Dual Sparsity through Activation-aware Pruning in LLMs
Ruokai Yin
Yuhang Li
Donghyun Lee
Priyadarshini Panda
VLM
190
2
0
25 Jun 2025
SLoPe: Double-Pruned Sparse Plus Lazy Low-Rank Adapter Pretraining of LLMs
SLoPe: Double-Pruned Sparse Plus Lazy Low-Rank Adapter Pretraining of LLMsInternational Conference on Learning Representations (ICLR), 2024
Mohammad Mozaffari
Amir Yazdanbakhsh
Zhao Zhang
M. Dehnavi
351
13
0
28 Jan 2025
1