ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2026 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2303.10464
  4. Cited By
SPDF: Sparse Pre-training and Dense Fine-tuning for Large Language
  Models
v1v2 (latest)

SPDF: Sparse Pre-training and Dense Fine-tuning for Large Language Models

Conference on Uncertainty in Artificial Intelligence (UAI), 2023
18 March 2023
Vithursan Thangarasa
Abhay Gupta
William Marshall
Tianda Li
Kevin Leong
D. DeCoste
Sean Lie
Shreyas Saxena
    MoEAI4CE
ArXiv (abs)PDFHTMLHuggingFace (1 upvotes)Github

Papers citing "SPDF: Sparse Pre-training and Dense Fine-tuning for Large Language Models"

12 / 12 papers shown
An Empirical Study of Sample Selection Strategies for Large Language Model Repair
An Empirical Study of Sample Selection Strategies for Large Language Model Repair
Xuran Li
Jingyi Wang
KELM
171
0
0
23 Oct 2025
Memory-Efficient Fine-Tuning via Low-Rank Activation Compression
Memory-Efficient Fine-Tuning via Low-Rank Activation Compression
Jiang-Xin Shi
Wen-Da Wei
Jin-Fei Qi
Xuanyu Chen
Tong Wei
Yu-Feng Li
176
0
0
27 Sep 2025
CR-Net: Scaling Parameter-Efficient Training with Cross-Layer Low-Rank Structure
CR-Net: Scaling Parameter-Efficient Training with Cross-Layer Low-Rank Structure
Boao Kong
Junzhu Liang
Yuxi Liu
Renjia Deng
Kun Yuan
220
2
0
23 Sep 2025
Generalizing Scaling Laws for Dense and Sparse Large Language Models
Generalizing Scaling Laws for Dense and Sparse Large Language Models
Md Arafat Hossain
Xingfu Wu
V. Taylor
Ali Jannesari
228
0
0
08 Aug 2025
SparseLoRA: Accelerating LLM Fine-Tuning with Contextual Sparsity
SparseLoRA: Accelerating LLM Fine-Tuning with Contextual Sparsity
Samir Khaki
Xiuyu Li
Junxian Guo
Ligeng Zhu
Chenfeng Xu
Konstantinos N. Plataniotis
Amir Yazdanbakhsh
Kurt Keutzer
Song Han
Zhijian Liu
253
7
0
19 Jun 2025
SD$^2$: Self-Distilled Sparse Drafters
SD2^22: Self-Distilled Sparse Drafters
Mike Lasby
Nish Sinnadurai
Valavan Manohararajah
Sean Lie
Yani Andrew Ioannou
Vithursan Thangarasa
846
1
0
10 Apr 2025
Communication-Efficient and Personalized Federated Foundation Model Fine-Tuning via Tri-Matrix Adaptation
Communication-Efficient and Personalized Federated Foundation Model Fine-Tuning via Tri-Matrix Adaptation
Yongqian Li
Bo Liu
Sheng Huang
Zhe Zhang
Xiaotong Yuan
Richang Hong
502
1
0
31 Mar 2025
The impact of allocation strategies in subset learning on the expressive power of neural networks
The impact of allocation strategies in subset learning on the expressive power of neural networksInternational Conference on Learning Representations (ICLR), 2025
Ofir Schlisselberg
Ran Darshan
355
0
0
10 Feb 2025
Brain network science modelling of sparse neural networks enables Transformers and LLMs to perform as fully connected
Brain network science modelling of sparse neural networks enables Transformers and LLMs to perform as fully connected
Yingtao Zhang
Diego Cerretti
Jialin Zhao
Wenjing Wu
Ziheng Liao
Umberto Michieli
C. Cannistraci
733
1
0
31 Jan 2025
SLoPe: Double-Pruned Sparse Plus Lazy Low-Rank Adapter Pretraining of LLMs
SLoPe: Double-Pruned Sparse Plus Lazy Low-Rank Adapter Pretraining of LLMsInternational Conference on Learning Representations (ICLR), 2024
Mohammad Mozaffari
Amir Yazdanbakhsh
Zhao Zhang
M. Dehnavi
495
15
0
28 Jan 2025
Enhancing Zero-Shot Crypto Sentiment with Fine-tuned Language Model and
  Prompt Engineering
Enhancing Zero-Shot Crypto Sentiment with Fine-tuned Language Model and Prompt Engineering
Rahman S. M. Wahidur
Ishmam Tashdeed
Manjit Kaur
Heung-No Lee
ALM
324
29
0
20 Oct 2023
Sparsifying Spiking Networks through Local Rhythms
Sparsifying Spiking Networks through Local RhythmsInternational Conference on Systems (ICONS), 2023
W. Olin-Ammentorp
171
0
0
30 Apr 2023
1
Page 1 of 1