ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2026 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2102.01732
  4. Cited By
Truly Sparse Neural Networks at Scale
v1v2 (latest)

Truly Sparse Neural Networks at Scale

2 February 2021
Selima Curci
Decebal Constantin Mocanu
Mykola Pechenizkiy
ArXiv (abs)PDFHTML

Papers citing "Truly Sparse Neural Networks at Scale"

16 / 16 papers shown
NeuroTrails: Training with Dynamic Sparse Heads as the Key to Effective Ensembling
NeuroTrails: Training with Dynamic Sparse Heads as the Key to Effective Ensembling
Bram Grooten
Farid Hasanov
Chenxiang Zhang
Q. Xiao
Boqian Wu
...
Shiwei Liu
L. Yin
Elena Mocanu
Mykola Pechenizkiy
Decebal Constantin Mocanu
415
0
0
23 May 2025
Sparse-to-Sparse Training of Diffusion Models
Sparse-to-Sparse Training of Diffusion Models
Inês Cardoso Oliveira
Decebal Constantin Mocanu
Luis A. Leiva
DiffM
602
1
0
30 Apr 2025
Navigating Extremes: Dynamic Sparsity in Large Output Spaces
Navigating Extremes: Dynamic Sparsity in Large Output SpacesNeural Information Processing Systems (NeurIPS), 2024
Nasib Ullah
Erik Schultheis
Mike Lasby
Yani Andrew Ioannou
Rohit Babbar
514
3
0
05 Nov 2024
Are Sparse Neural Networks Better Hard Sample Learners?
Are Sparse Neural Networks Better Hard Sample Learners?British Machine Vision Conference (BMVC), 2024
Q. Xiao
Boqian Wu
Lu Yin
Christopher Neil Gadzinski
Tianjin Huang
Mykola Pechenizkiy
Decebal Constantin Mocanu
270
2
0
13 Sep 2024
Nerva: a Truly Sparse Implementation of Neural Networks
Nerva: a Truly Sparse Implementation of Neural Networks
Wieger Wesselink
Bram Grooten
Qiao Xiao
Cássio Machado de Campos
Mykola Pechenizkiy
200
4
0
24 Jul 2024
Exploring the Relationship: Transformative Adaptive Activation Functions
  in Comparison to Other Activation Functions
Exploring the Relationship: Transformative Adaptive Activation Functions in Comparison to Other Activation Functions
Vladimír Kunc
443
3
0
14 Feb 2024
Fantastic Weights and How to Find Them: Where to Prune in Dynamic Sparse
  Training
Fantastic Weights and How to Find Them: Where to Prune in Dynamic Sparse TrainingNeural Information Processing Systems (NeurIPS), 2023
A. Nowak
Bram Grooten
Decebal Constantin Mocanu
Jacek Tabor
336
16
0
21 Jun 2023
Adaptive Sparsity Level during Training for Efficient Time Series
  Forecasting with Transformers
Adaptive Sparsity Level during Training for Efficient Time Series Forecasting with Transformers
Zahra Atashgahi
Mykola Pechenizkiy
Raymond N. J. Veldhuis
Decebal Constantin Mocanu
AI4TSAI4CE
342
2
0
28 May 2023
Learning Activation Functions for Sparse Neural Networks
Learning Activation Functions for Sparse Neural Networks
Mohammad Loni
Aditya Mohan
Mehdi Asadi
Marius Lindauer
290
6
0
18 May 2023
Automatic Noise Filtering with Dynamic Sparse Training in Deep
  Reinforcement Learning
Automatic Noise Filtering with Dynamic Sparse Training in Deep Reinforcement LearningAdaptive Agents and Multi-Agent Systems (AAMAS), 2023
Bram Grooten
Ghada Sokar
Shibhansh Dohare
Elena Mocanu
Matthew E. Taylor
Mykola Pechenizkiy
Decebal Constantin Mocanu
242
16
0
13 Feb 2023
Ten Lessons We Have Learned in the New "Sparseland": A Short Handbook
  for Sparse Neural Network Researchers
Ten Lessons We Have Learned in the New "Sparseland": A Short Handbook for Sparse Neural Network Researchers
Shiwei Liu
Zinan Lin
513
37
0
06 Feb 2023
Where to Pay Attention in Sparse Training for Feature Selection?
Where to Pay Attention in Sparse Training for Feature Selection?Neural Information Processing Systems (NeurIPS), 2022
Ghada Sokar
Zahra Atashgahi
Mykola Pechenizkiy
Decebal Constantin Mocanu
348
24
0
26 Nov 2022
Don't Be So Dense: Sparse-to-Sparse GAN Training Without Sacrificing
  Performance
Don't Be So Dense: Sparse-to-Sparse GAN Training Without Sacrificing PerformanceInternational Journal of Computer Vision (IJCV), 2022
Shiwei Liu
Yuesong Tian
Tianlong Chen
Li Shen
344
14
0
05 Mar 2022
A phase transition for finding needles in nonlinear haystacks with LASSO
  artificial neural networks
A phase transition for finding needles in nonlinear haystacks with LASSO artificial neural networksStatistics and computing (Stat. Comput.), 2022
Xiaoyu Ma
S. Sardy
N. Hengartner
Nikolai Bobenko
Yen Ting Lin
232
3
0
21 Jan 2022
Stochastic Spatio-Temporal Optimization for Control and Co-Design of
  Systems in Robotics and Applied Physics
Stochastic Spatio-Temporal Optimization for Control and Co-Design of Systems in Robotics and Applied PhysicsAutonomous Robots (Auton. Robots), 2021
Ethan N. Evans
Andrew P. Kendall
Evangelos A. Theodorou
AI4CE
283
11
0
18 Feb 2021
Gradient Flow in Sparse Neural Networks and How Lottery Tickets Win
Gradient Flow in Sparse Neural Networks and How Lottery Tickets Win
Utku Evci
Yani Andrew Ioannou
Cem Keskin
Yann N. Dauphin
269
106
0
07 Oct 2020
1
Page 1 of 1