ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2026 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1803.03635
  4. Cited By
The Lottery Ticket Hypothesis: Finding Sparse, Trainable Neural Networks
v1v2v3v4v5 (latest)

The Lottery Ticket Hypothesis: Finding Sparse, Trainable Neural Networks

9 March 2018
Jonathan Frankle
Michael Carbin
ArXiv (abs)PDFHTML

Papers citing "The Lottery Ticket Hypothesis: Finding Sparse, Trainable Neural Networks"

50 / 2,186 papers shown
SQS: Bayesian DNN Compression through Sparse Quantized Sub-distributions
SQS: Bayesian DNN Compression through Sparse Quantized Sub-distributions
Ziyi Wang
Nan Jiang
Guang Lin
Qifan Song
MQ
209
0
0
10 Oct 2025
Robustness and Regularization in Hierarchical Re-Basin
Robustness and Regularization in Hierarchical Re-BasinThe European Symposium on Artificial Neural Networks (ESANN), 2025
Benedikt Franke
Florian Heinrich
Markus Lange
Arne P. Raulf
MoMeAAML
275
1
0
10 Oct 2025
Do We Really Need Permutations? Impact of Width Expansion on Linear Mode Connectivity
Do We Really Need Permutations? Impact of Width Expansion on Linear Mode Connectivity
Akira Ito
Masanori Yamada
Daiki Chijiwa
Atsutoshi Kumagai
MoMe
190
0
0
09 Oct 2025
SliceFine: The Universal Winning-Slice Hypothesis for Pretrained Networks
SliceFine: The Universal Winning-Slice Hypothesis for Pretrained Networks
Md. Kowsher
Ali O. Polat
Ehsan Mohammady Ardehaly
Mehrdad Salehi
Zia Ghiasi
Prasanth Murali
Chen Chen
186
2
0
09 Oct 2025
JAI-1: A Thai-Centric Large Language Model
JAI-1: A Thai-Centric Large Language Model
Attapol T. Rutherford
Jullajak Karnjanaekarin
Narongkorn Panitsrisit
Pontakorn Trakuekul
Sumana Sumanakul
Natchanon Pollertlam
82
0
0
08 Oct 2025
Where to Begin: Efficient Pretraining via Subnetwork Selection and Distillation
Where to Begin: Efficient Pretraining via Subnetwork Selection and Distillation
Arjun Krishnakumar
R. Sukthanker
Hannan Javed Mahadik
Gabriela Kadlecová
Vladyslav Moroshan
Timur Carstensen
Frank Hutter
Aaron Klein
132
0
0
08 Oct 2025
GUIDE: Guided Initialization and Distillation of Embeddings
GUIDE: Guided Initialization and Distillation of Embeddings
Khoa Trinh
Gaurav Menghani
Erik Vee
122
0
0
07 Oct 2025
Diversity Is All You Need for Contrastive Learning: Spectral Bounds on Gradient Magnitudes
Diversity Is All You Need for Contrastive Learning: Spectral Bounds on Gradient Magnitudes
Peter Ochieng
88
1
0
07 Oct 2025
Expand Neurons, Not Parameters
Expand Neurons, Not Parameters
Linghao Kong
Inimai Subramanian
Yonadav Shavit
Micah Adler
Dan Alistarh
Nir Shavit
127
0
0
06 Oct 2025
PolyKAN: A Polyhedral Analysis Framework for Provable and Approximately Optimal KAN Compression
PolyKAN: A Polyhedral Analysis Framework for Provable and Approximately Optimal KAN Compression
Di Zhang
124
0
0
05 Oct 2025
Categorical Invariants of Learning Dynamics
Categorical Invariants of Learning Dynamics
Abdulrahman Tamim
OOD
116
0
0
05 Oct 2025
CHORD: Customizing Hybrid-precision On-device Model for Sequential Recommendation with Device-cloud Collaboration
CHORD: Customizing Hybrid-precision On-device Model for Sequential Recommendation with Device-cloud Collaboration
Tianqi Liu
Kairui Fu
Shengyu Zhang
W. Fan
Zhaocheng Du
Jieming Zhu
Fan Wu
Fei Wu
MQ
123
0
0
03 Oct 2025
FlexiQ: Adaptive Mixed-Precision Quantization for Latency/Accuracy Trade-Offs in Deep Neural Networks
FlexiQ: Adaptive Mixed-Precision Quantization for Latency/Accuracy Trade-Offs in Deep Neural Networks
Jaemin Kim
Hongjun Um
Sungkyun Kim
Yongjun Park
Jiwon Seo
MQ
143
0
0
03 Oct 2025
Shift-Invariant Attribute Scoring for Kolmogorov-Arnold Networks via Shapley Value
Shift-Invariant Attribute Scoring for Kolmogorov-Arnold Networks via Shapley Value
Wangxuan Fan
Ching Wang
Siqi Li
Nan Liu
FAtt
218
1
0
02 Oct 2025
The Unseen Frontier: Pushing the Limits of LLM Sparsity with Surrogate-Free ADMM
The Unseen Frontier: Pushing the Limits of LLM Sparsity with Surrogate-Free ADMM
Kwanhee Lee
Hyeondo Jang
Dongyeop Lee
Dan Alistarh
Namhoon Lee
95
1
0
02 Oct 2025
A universal compression theory: Lottery ticket hypothesis and superpolynomial scaling laws
A universal compression theory: Lottery ticket hypothesis and superpolynomial scaling laws
Hong-Yi Wang
Di Luo
T. Poggio
Isaac Chuang
Liu Ziyin
81
1
0
01 Oct 2025
Interpret, prune and distill Donut : towards lightweight VLMs for VQA on document
Interpret, prune and distill Donut : towards lightweight VLMs for VQA on document
Adnan Ben Mansour
Ayoub Karine
D. Naccache
130
0
0
30 Sep 2025
Effective Model Pruning
Effective Model Pruning
Yixuan Wang
Dan Guralnik
Saiedeh Akbari
Warren E. Dixon
53
0
0
30 Sep 2025
Enhancing Certifiable Semantic Robustness via Robust Pruning of Deep Neural Networks
Enhancing Certifiable Semantic Robustness via Robust Pruning of Deep Neural Networks
Hanjiang Hu
Bowei Li
Ziwei Wang
Tianhao Wei
Casidhe Hutchison
Eric Sample
Changliu Liu
AAML
135
0
0
30 Sep 2025
Growing Winning Subnetworks, Not Pruning Them: A Paradigm for Density Discovery in Sparse Neural Networks
Growing Winning Subnetworks, Not Pruning Them: A Paradigm for Density Discovery in Sparse Neural Networks
Qihang Yao
Constantine Dovrolis
120
0
0
30 Sep 2025
CAST: Continuous and Differentiable Semi-Structured Sparsity-Aware Training for Large Language Models
CAST: Continuous and Differentiable Semi-Structured Sparsity-Aware Training for Large Language Models
Weiyu Huang
Yuezhou Hu
Jun Zhu
Jianfei Chen
CLL
108
0
0
30 Sep 2025
Query Circuits: Explaining How Language Models Answer User Prompts
Query Circuits: Explaining How Language Models Answer User Prompts
Tung-Yu Wu
Fazl Barez
ReLMLRM
154
0
0
29 Sep 2025
How LLMs Learn to Reason: A Complex Network Perspective
How LLMs Learn to Reason: A Complex Network Perspective
Sihan Hu
X-D Cai
Yuan Huang
Zhiyuan Yao
Linfeng Zhang
Pan Zhang
Youjin Deng
Kun Chen
LRM
232
1
0
28 Sep 2025
Differentiable Sparsity via $D$-Gating: Simple and Versatile Structured Penalization
Differentiable Sparsity via DDD-Gating: Simple and Versatile Structured Penalization
Chris Kolb
Laetitia Frost
J. Herbinger
David Rügamer
383
0
0
28 Sep 2025
A Second-Order Perspective on Pruning at Initialization and Knowledge Transfer
A Second-Order Perspective on Pruning at Initialization and Knowledge Transfer
Leonardo Iurada
Beatrice Occhiena
Tatiana Tommasi
VLM
154
0
0
28 Sep 2025
Evaluating the Robustness of Chinchilla Compute-Optimal Scaling
Evaluating the Robustness of Chinchilla Compute-Optimal Scaling
Rylan Schaeffer
Noam Levi
Andreas Kirsch
Theo Guenais
Brando Miranda
Elyas Obbad
Sanmi Koyejo
LRM
187
0
0
28 Sep 2025
MonoCon: A general framework for learning ultra-compact high-fidelity representations using monotonicity constraints
MonoCon: A general framework for learning ultra-compact high-fidelity representations using monotonicity constraints
Shreyas Gokhale
112
0
0
26 Sep 2025
COSPADI: Compressing LLMs via Calibration-Guided Sparse Dictionary Learning
COSPADI: Compressing LLMs via Calibration-Guided Sparse Dictionary Learning
Dmitriy Shopkhoev
Denis Makhov
Magauiya Zhussip
Ammar Ali
Stamatios Lefkimmiatis
186
0
0
26 Sep 2025
Budgeted Broadcast: An Activity-Dependent Pruning Rule for Neural Network Efficiency
Budgeted Broadcast: An Activity-Dependent Pruning Rule for Neural Network Efficiency
Yaron Meirovitch
Fuming Yang
J. Lichtman
Nir Shavit
111
1
0
26 Sep 2025
Toward a Theory of Generalizability in LLM Mechanistic Interpretability Research
Toward a Theory of Generalizability in LLM Mechanistic Interpretability Research
Sean Trott
112
1
0
26 Sep 2025
Can Less Precise Be More Reliable? A Systematic Evaluation of Quantization's Impact on CLIP Beyond Accuracy
Can Less Precise Be More Reliable? A Systematic Evaluation of Quantization's Impact on CLIP Beyond Accuracy
Aymen Bouguerra
Daniel Montoya
Alexandra Gomez-Villa
Fabio Arnez
Chokri Mraidha
UQCV
327
0
0
25 Sep 2025
Binary Autoencoder for Mechanistic Interpretability of Large Language Models
Binary Autoencoder for Mechanistic Interpretability of Large Language Models
Hakaze Cho
Haolin Yang
Brian M. Kurkoski
Naoya Inoue
MQ
200
0
0
25 Sep 2025
Smaller is Better: Enhancing Transparency in Vehicle AI Systems via Pruning
Smaller is Better: Enhancing Transparency in Vehicle AI Systems via Pruning
Sanish Suwal
Shaurya Garg
Dipkamal Bhusal
Michael Clifford
Nidhi Rastogi
AAML
145
1
0
24 Sep 2025
A Recovery Guarantee for Sparse Neural Networks
A Recovery Guarantee for Sparse Neural Networks
Sara Fridovich-Keil
Mert Pilanci
86
0
0
24 Sep 2025
Do Sparse Subnetworks Exhibit Cognitively Aligned Attention? Effects of Pruning on Saliency Map Fidelity, Sparsity, and Concept Coherence
Do Sparse Subnetworks Exhibit Cognitively Aligned Attention? Effects of Pruning on Saliency Map Fidelity, Sparsity, and Concept Coherence
Sanish Suwal
Dipkamal Bhusal
Michael Clifford
Nidhi Rastogi
258
1
0
23 Sep 2025
Efficient Reinforcement Learning by Reducing Forgetting with Elephant Activation Functions
Efficient Reinforcement Learning by Reducing Forgetting with Elephant Activation Functions
Qingfeng Lan
Gautham Vasan
A. R. Mahmood
CLL
128
0
0
23 Sep 2025
TASO: Task-Aligned Sparse Optimization for Parameter-Efficient Model Adaptation
TASO: Task-Aligned Sparse Optimization for Parameter-Efficient Model Adaptation
Daiye Miao
Yufang Liu
Jie Wang
Changzhi Sun
Yunke Zhang
Demei Yan
Shaokang Dong
Qi Zhang
Man Lan
87
0
0
22 Sep 2025
A Novel Differential Feature Learning for Effective Hallucination Detection and Classification
A Novel Differential Feature Learning for Effective Hallucination Detection and Classification
Wenkai Wang
Vincent C. S. Lee
Yizhen Zheng
86
0
0
20 Sep 2025
Analyzing the Effects of Supervised Fine-Tuning on Model Knowledge from Token and Parameter Levels
Analyzing the Effects of Supervised Fine-Tuning on Model Knowledge from Token and Parameter Levels
Junjie Ye
Yuming Yang
Yang Nan
Shuo Li
Qi Zhang
Tao Gui
Xuanjing Huang
Liang Luo
Zhongchao Shi
Jianping Fan
115
1
0
20 Sep 2025
CoUn: Empowering Machine Unlearning via Contrastive Learning
CoUn: Empowering Machine Unlearning via Contrastive Learning
Yasser H. Khalil
Mehdi Setayesh
Hongliang Li
MUCLL
269
0
0
19 Sep 2025
Towards Privacy-Preserving and Heterogeneity-aware Split Federated Learning via Probabilistic Masking
Towards Privacy-Preserving and Heterogeneity-aware Split Federated Learning via Probabilistic Masking
Xingchen Wang
Feijie Wu
Chenglin Miao
Tianchun Li
Haoyu Hu
Qiming Cao
Jing Gao
Lu Su
160
0
0
18 Sep 2025
Balancing Sparse RNNs with Hyperparameterization Benefiting Meta-Learning
Balancing Sparse RNNs with Hyperparameterization Benefiting Meta-Learning
Quincy Hershey
Randy Paffenroth
44
0
0
18 Sep 2025
Deep Learning-Driven Peptide Classification in Biological Nanopores
Deep Learning-Driven Peptide Classification in Biological Nanopores
S. Tovey
Julian Hoßbach
Sandro Kuppel
Tobias Ensslen
Jan C. Behrends
Christian Holm
114
0
0
17 Sep 2025
NIRVANA: Structured pruning reimagined for large language models compression
NIRVANA: Structured pruning reimagined for large language models compression
Mengting Ai
Tianxin Wei
Sirui Chen
Jingrui He
VLM
1.6K
1
0
17 Sep 2025
Spontaneous Kolmogorov-Arnold Geometry in Shallow MLPs
Spontaneous Kolmogorov-Arnold Geometry in Shallow MLPs
Michael Freedman
Michael Mulligan
56
1
0
15 Sep 2025
NeuroStrike: Neuron-Level Attacks on Aligned LLMs
NeuroStrike: Neuron-Level Attacks on Aligned LLMs
Lichao Wu
Sasha Behrouzi
Mohamadreza Rostami
Maximilian Thang
S. Picek
A. Sadeghi
AAML
237
1
0
15 Sep 2025
Investigating the Lottery Ticket Hypothesis for Variational Quantum Circuits
Investigating the Lottery Ticket Hypothesis for Variational Quantum Circuits
Michael Kölle
Leonhard Klingert
Julian Schonberger
Philipp Altmann
Tobias Rohe
Claudia Linnhoff-Popien
105
0
0
14 Sep 2025
Optimal Brain Restoration for Joint Quantization and Sparsification of LLMs
Optimal Brain Restoration for Joint Quantization and Sparsification of LLMs
Hang Guo
Yawei Li
Luca Benini
MQ
211
0
0
14 Sep 2025
Explaining How Quantization Disparately Skews a Model
Explaining How Quantization Disparately Skews a Model
Abhimanyu Bellam
Jung-Eun Kim
MQ
183
0
0
08 Sep 2025
Robust Experts: the Effect of Adversarial Training on CNNs with Sparse Mixture-of-Experts Layers
Robust Experts: the Effect of Adversarial Training on CNNs with Sparse Mixture-of-Experts Layers
Svetlana Pavlitska
Haixi Fan
Konstantin Ditschuneit
Johann Marius Zöllner
AAMLMoE
98
0
0
05 Sep 2025
Previous
12345...424344
Next