ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2026 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1803.03635
  4. Cited By
The Lottery Ticket Hypothesis: Finding Sparse, Trainable Neural Networks
v1v2v3v4v5 (latest)

The Lottery Ticket Hypothesis: Finding Sparse, Trainable Neural Networks

9 March 2018
Jonathan Frankle
Michael Carbin
ArXiv (abs)PDFHTML

Papers citing "The Lottery Ticket Hypothesis: Finding Sparse, Trainable Neural Networks"

50 / 2,186 papers shown
Sharp Generalization Bounds for Foundation Models with Asymmetric Randomized Low-Rank Adapters
Sharp Generalization Bounds for Foundation Models with Asymmetric Randomized Low-Rank Adapters
Anastasis Kratsios
Tin Sum Cheng
Aurelien Lucchi
Haitz Sáez de Ocáriz Borde
210
1
0
17 Jun 2025
Dynamic Acoustic Model Architecture Optimization in Training for ASR
Dynamic Acoustic Model Architecture Optimization in Training for ASR
Jingjing Xu
Zijian Yang
Albert Zeyer
Eugen Beck
Ralf Schlueter
Hermann Ney
215
0
0
16 Jun 2025
The Butterfly Effect: Neural Network Training Trajectories Are Highly Sensitive to Initial Conditions
The Butterfly Effect: Neural Network Training Trajectories Are Highly Sensitive to Initial Conditions
Devin Kwok
Gül Sena Altıntaş
Colin Raffel
David Rolnick
407
2
0
16 Jun 2025
MaskPro: Linear-Space Probabilistic Learning for Strict (N:M)-Sparsity on Large Language Models
MaskPro: Linear-Space Probabilistic Learning for Strict (N:M)-Sparsity on Large Language Models
Yan Sun
Qixin Zhang
Zhiyuan Yu
Xikun Zhang
Li Shen
Dacheng Tao
199
1
0
15 Jun 2025
Compression Aware Certified Training
Compression Aware Certified Training
Changming Xu
Gagandeep Singh
162
0
0
13 Jun 2025
Auto-Compressing Networks
Auto-Compressing Networks
Vaggelis Dorovatas
Georgios Paraskevopoulos
Alexandros Potamianos
393
2
0
11 Jun 2025
The Emergence of Abstract Thought in Large Language Models Beyond Any Language
The Emergence of Abstract Thought in Large Language Models Beyond Any Language
Yuxin Chen
Yiran Zhao
Yang Zhang
An Zhang
Kenji Kawaguchi
Shafiq Joty
Junnan Li
Tat-Seng Chua
Michael Shieh
Wenxuan Zhang
LRM
185
6
0
11 Jun 2025
Olica: Efficient Structured Pruning of Large Language Models without Retraining
Jiujun He
Huazhen Lin
172
1
0
10 Jun 2025
On the Stability of the Jacobian Matrix in Deep Neural Networks
Benjamin Dadoun
Soufiane Hayou
Hanan Salam
Abdalgader Abubaker
Pierre Youssef
ODL
387
0
0
10 Jun 2025
Hyperpruning: Efficient Search through Pruned Variants of Recurrent Neural Networks Leveraging Lyapunov Spectrum
Hyperpruning: Efficient Search through Pruned Variants of Recurrent Neural Networks Leveraging Lyapunov Spectrum
Caleb Zheng
Eli Shlizerman
166
0
0
09 Jun 2025
PrunePEFT: Iterative Hybrid Pruning for Parameter-Efficient Fine-tuning of LLMs
PrunePEFT: Iterative Hybrid Pruning for Parameter-Efficient Fine-tuning of LLMs
Tongzhou Yu
Zhuhao Zhang
Guanghui Zhu
Shen Jiang
Meikang Qiu
Yihua Huang
152
1
0
09 Jun 2025
Deep RL Needs Deep Behavior Analysis: Exploring Implicit Planning by Model-Free Agents in Open-Ended Environments
Deep RL Needs Deep Behavior Analysis: Exploring Implicit Planning by Model-Free Agents in Open-Ended Environments
Riley Simmons-Edler
R. Badman
Felix Baastad Berg
Raymond Chua
John J. Vastola
Joshua Lunger
William Qian
K. Rajan
OffRL
239
3
0
08 Jun 2025
SAFE: Finding Sparse and Flat Minima to Improve Pruning
SAFE: Finding Sparse and Flat Minima to Improve Pruning
Dongyeop Lee
Kwanhee Lee
Jinseok Chung
Namhoon Lee
293
4
0
07 Jun 2025
Neural Network Reprogrammability: A Unified Theme on Model Reprogramming, Prompt Tuning, and Prompt Instruction
Neural Network Reprogrammability: A Unified Theme on Model Reprogramming, Prompt Tuning, and Prompt Instruction
Zesheng Ye
C. Cai
Ruijiang Dong
Jianzhong Qi
Bingquan Shen
Pin-Yu Chen
Feng Liu
649
1
0
05 Jun 2025
Weisfeiler and Leman Go Gambling: Why Expressive Lottery Tickets Win
Weisfeiler and Leman Go Gambling: Why Expressive Lottery Tickets Win
Lorenz Kummer
Samir Moustafa
Anatol Ehrlich
Franka Bause
Nikolaus Suess
Wilfried Gansterer
Nils M. Kriege
208
0
0
04 Jun 2025
HAM: A Hyperbolic Step to Regulate Implicit Bias
HAM: A Hyperbolic Step to Regulate Implicit Bias
Tom Jacobs
Advait Gadhikar
Celia Rubio-Madrigal
R. Burkholz
217
0
0
03 Jun 2025
Brain-Like Processing Pathways Form in Models With Heterogeneous Experts
Brain-Like Processing Pathways Form in Models With Heterogeneous Experts
Jack Cook
Danyal Akarca
Rui Ponte Costa
Jascha Achterberg
MoE
237
3
0
03 Jun 2025
Computation- and Communication-Efficient Online FL for Resource-Constrained Aerial Vehicles
Computation- and Communication-Efficient Online FL for Resource-Constrained Aerial Vehicles
Md Ferdous Pervej
Richeng Jin
Md Moin Uddin Chowdhury
Simran Singh
Ismail Güvenç
H. Dai
213
0
0
03 Jun 2025
The Future of Continual Learning in the Era of Foundation Models: Three Key Directions
The Future of Continual Learning in the Era of Foundation Models: Three Key Directions
Jack Bell
Luigi Quarantiello
Eric Nuertey Coleman
Lanpei Li
Malio Li
Mauro Madeddu
Elia Piccoli
Vincenzo Lomonaco
KELM
260
5
0
03 Jun 2025
Frugal Machine Learning for Energy-efficient, and Resource-aware Artificial Intelligence
Frugal Machine Learning for Energy-efficient, and Resource-aware Artificial Intelligence
John Violos
Konstantina-Christina Diamanti
I. Kompatsiaris
Symeon Papadopoulos
210
1
0
02 Jun 2025
Smooth Model Compression without Fine-Tuning
Smooth Model Compression without Fine-Tuning
Christina Runkel
Natacha Kuete Meli
Jovita Lukasik
A. Biguri
Carola-Bibiane Schönlieb
Michael Moeller
245
0
0
30 May 2025
Learning Interpretable Differentiable Logic Networks for Tabular Regression
Learning Interpretable Differentiable Logic Networks for Tabular Regression
C. Yue
N. Jha
352
1
0
29 May 2025
ACE: Exploring Activation Cosine Similarity and Variance for Accurate and Calibration-Efficient LLM Pruning
ACE: Exploring Activation Cosine Similarity and Variance for Accurate and Calibration-Efficient LLM Pruning
Zhendong Mi
Zhenglun Kong
Geng Yuan
Shaoyi Huang
239
2
0
28 May 2025
M-Wanda: Improving One-Shot Pruning for Multilingual LLMs
M-Wanda: Improving One-Shot Pruning for Multilingual LLMs
Rochelle Choenni
Ivan Titov
219
1
0
27 May 2025
Sparsified State-Space Models are Efficient Highway Networks
Sparsified State-Space Models are Efficient Highway Networks
Woomin Song
Jihoon Tack
Sangwoo Mo
Seunghyuk Oh
Jinwoo Shin
Mamba
340
0
0
27 May 2025
Global Minimizers of $\ell^p$-Regularized Objectives Yield the Sparsest ReLU Neural Networks
Global Minimizers of ℓp\ell^pℓp-Regularized Objectives Yield the Sparsest ReLU Neural Networks
Julia B. Nakhleh
Robert D. Nowak
264
0
0
27 May 2025
DLP: Dynamic Layerwise Pruning in Large Language Models
DLP: Dynamic Layerwise Pruning in Large Language Models
Yuli Chen
B. Cheng
Jiale Han
Yingying Zhang
Yingting Li
Shuhao Zhang
258
1
0
27 May 2025
Binarized Neural Networks Converge Toward Algorithmic Simplicity: Empirical Support for the Learning-as-Compression Hypothesis
Binarized Neural Networks Converge Toward Algorithmic Simplicity: Empirical Support for the Learning-as-Compression Hypothesis
Eduardo Y. Sakabe
Felipe S. Abrahão
A. S. Simões
Esther L. Colombini
P. Costa
Ricardo Ribeiro Gudwin
Hector Zenil
376
0
0
27 May 2025
WINA: Weight Informed Neuron Activation for Accelerating Large Language Model Inference
WINA: Weight Informed Neuron Activation for Accelerating Large Language Model Inference
Sihan Chen
Dan Zhao
Jongwoo Ko
Colby R. Banbury
Huiping Zhuang
Luming Liang
Tianyi Chen
177
0
0
26 May 2025
GraSS: Scalable Data Attribution with Gradient Sparsification and Sparse Projection
GraSS: Scalable Data Attribution with Gradient Sparsification and Sparse Projection
Pingbang Hu
Joseph Melkonian
Weijing Tang
Han Zhao
Jiaqi W. Ma
TDI
577
0
0
25 May 2025
PacTrain: Pruning and Adaptive Sparse Gradient Compression for Efficient Collective Communication in Distributed Deep Learning
PacTrain: Pruning and Adaptive Sparse Gradient Compression for Efficient Collective Communication in Distributed Deep LearningDesign Automation Conference (DAC), 2025
Yisu Wang
Ruilong Wu
Xinjiao Li
Dirk Kutscher
355
0
0
24 May 2025
$μ$-MoE: Test-Time Pruning as Micro-Grained Mixture-of-Experts
μμμ-MoE: Test-Time Pruning as Micro-Grained Mixture-of-Experts
T. Koike-Akino
Jing Liu
Ye Wang
MoE
213
0
0
24 May 2025
Learning without Isolation: Pathway Protection for Continual Learning
Learning without Isolation: Pathway Protection for Continual Learning
Zhikang Chen
Abudukelimu Wuerkaixi
Sen Cui
Haoxuan Li
D. Li
...
Houfang Liu
Yi Yang
Sifan Yang
Changshui Zhang
Tianling Ren
CLL
187
2
0
24 May 2025
Meta Pruning via Graph Metanetworks : A Universal Meta Learning Framework for Network Pruning
Meta Pruning via Graph Metanetworks : A Universal Meta Learning Framework for Network Pruning
Yewei Liu
Xiyuan Wang
Muhan Zhang
DDGNN
345
0
0
24 May 2025
How Many Parameters Does Your Task Really Need? Task Specific Pruning with LLM-Sieve
How Many Parameters Does Your Task Really Need? Task Specific Pruning with LLM-Sieve
Waleed Reda
Abhinav Jangda
Krishna Chintalapudi
295
0
0
23 May 2025
The Unreasonable Effectiveness of Model Merging for Cross-Lingual Transfer in LLMs
The Unreasonable Effectiveness of Model Merging for Cross-Lingual Transfer in LLMs
Lucas Bandarkar
Nanyun Peng
MoMeLRM
308
1
0
23 May 2025
Structured Linear CDEs: Maximally Expressive and Parallel-in-Time Sequence Models
Structured Linear CDEs: Maximally Expressive and Parallel-in-Time Sequence Models
Benjamin Walker
Lingyi Yang
Nicola Muca Cirone
C. Salvi
Terry Lyons
AI4TS
362
6
0
23 May 2025
Generalized Fisher-Weighted SVD: Scalable Kronecker-Factored Fisher Approximation for Compressing Large Language Models
Generalized Fisher-Weighted SVD: Scalable Kronecker-Factored Fisher Approximation for Compressing Large Language Models
Viktoriia Chekalina
Daniil Moskovskiy
Daria Cherniuk
Maxim Kurkin
Andrey Kuznetsov
Evgeny Frolov
415
1
0
23 May 2025
PreMoe: Lightening MoEs on Constrained Memory by Expert Pruning and Retrieval
PreMoe: Lightening MoEs on Constrained Memory by Expert Pruning and Retrieval
Zehua Pei
Ying Zhang
Hui-Ling Zhen
Xianzhi Yu
Wulong Liu
Sinno Jialin Pan
Mingxuan Yuan
Bei Yu
MoE
172
0
0
23 May 2025
Model Editing with Graph-Based External Memory
Model Editing with Graph-Based External Memory
Yash Kumar Atri
Ahmed Alaa
Thomas Hartvigsen
KELM
204
0
0
23 May 2025
Locate-then-Merge: Neuron-Level Parameter Fusion for Mitigating Catastrophic Forgetting in Multimodal LLMs
Locate-then-Merge: Neuron-Level Parameter Fusion for Mitigating Catastrophic Forgetting in Multimodal LLMs
Zeping Yu
Sophia Ananiadou
MoMeKELMCLL
254
1
0
22 May 2025
TRIM: Achieving Extreme Sparsity with Targeted Row-wise Iterative Metric-driven Pruning
TRIM: Achieving Extreme Sparsity with Targeted Row-wise Iterative Metric-driven Pruning
Florentin Beck
William Rudman
Carsten Eickhoff
374
1
0
22 May 2025
DeFTX: Denoised Sparse Fine-Tuning for Zero-Shot Cross-Lingual Transfer
DeFTX: Denoised Sparse Fine-Tuning for Zero-Shot Cross-Lingual Transfer
Sona Elza Simon
Preethi Jyothi
VLM
305
1
0
21 May 2025
Balanced and Elastic End-to-end Training of Dynamic LLMs
Balanced and Elastic End-to-end Training of Dynamic LLMs
Mohamed Wahib
Muhammed Abdullah Soyturk
Didem Unat
MoE
327
1
0
20 May 2025
Reinforcement Learning Finetunes Small Subnetworks in Large Language Models
Reinforcement Learning Finetunes Small Subnetworks in Large Language Models
Sagnik Mukherjee
Lifan Yuan
Dilek Hakkani-Tur
Yuan Yao
283
14
0
16 May 2025
MID-L: Matrix-Interpolated Dropout Layer with Layer-wise Neuron Selection
MID-L: Matrix-Interpolated Dropout Layer with Layer-wise Neuron Selection
Pouya Shaeri
Ariane Middel
59
2
0
16 May 2025
Uniform Loss vs. Specialized Optimization: A Comparative Analysis in Multi-Task Learning
Uniform Loss vs. Specialized Optimization: A Comparative Analysis in Multi-Task Learning
Gabriel S. Gama
Valdir Grassi Jr
MoMe
301
0
0
15 May 2025
BINGO: A Novel Pruning Mechanism to Reduce the Size of Neural Networks
BINGO: A Novel Pruning Mechanism to Reduce the Size of Neural Networks
Aditya Panangat
260
0
0
15 May 2025
Are Spatial-Temporal Graph Convolution Networks for Human Action Recognition Over-Parameterized?
Are Spatial-Temporal Graph Convolution Networks for Human Action Recognition Over-Parameterized?Computer Vision and Pattern Recognition (CVPR), 2025
Jianyang Xie
Yitian Zhao
Y. Meng
He Zhao
Anh Nguyen
Yalin Zheng
247
3
0
15 May 2025
Low-Complexity Inference in Continual Learning via Compressed Knowledge Transfer
Low-Complexity Inference in Continual Learning via Compressed Knowledge Transfer
Zhenrong Liu
Janne M. J. Huttunen
Mikko Honkala
CLL
312
1
0
13 May 2025
Previous
12345...424344
Next