ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2405.09673
  4. Cited By
LoRA Learns Less and Forgets Less

LoRA Learns Less and Forgets Less

15 May 2024
D. Biderman
Jose Javier Gonzalez Ortiz
Jacob P. Portes
Mansheej Paul
Philip Greengard
Connor Jennings
Daniel King
Sam Havens
Vitaliy Chiley
Jonathan Frankle
Cody Blakeney
John P. Cunningham
    CLL
ArXiv (abs)PDFHTMLHuggingFace (89 upvotes)

Papers citing "LoRA Learns Less and Forgets Less"

50 / 141 papers shown
Title
Parameter Importance-Driven Continual Learning for Foundation Models
Parameter Importance-Driven Continual Learning for Foundation Models
LingXiang Wang
Hainan Zhang
Zhiming Zheng
KELMCLL
426
0
0
19 Nov 2025
On the Difficulty of Token-Level Modeling of Dysfluency and Fluency Shaping Artifacts
On the Difficulty of Token-Level Modeling of Dysfluency and Fluency Shaping Artifacts
Kashaf Gulzar
Dominik Wagner
Sebastian P. Bayerl
Florian Honig
Tobias Bocklet
Korbinian Riedhammer
12
0
0
18 Nov 2025
Random Initialization of Gated Sparse Adapters
Random Initialization of Gated Sparse Adapters
Vi Retault
Yohaï-Eliel Berreby
CLLMoE
184
0
0
03 Nov 2025
A Comparative Analysis of LLM Adaptation: SFT, LoRA, and ICL in Data-Scarce Scenarios
A Comparative Analysis of LLM Adaptation: SFT, LoRA, and ICL in Data-Scarce Scenarios
Bernd Bohnet
Rumen Dangovski
Kevin Swersky
Sherry Moore
Arslan Chaudhry
Kathleen Kenealy
Noah Fiedel
CLL
254
0
0
31 Oct 2025
LoRAQuant: Mixed-Precision Quantization of LoRA to Ultra-Low Bits
LoRAQuant: Mixed-Precision Quantization of LoRA to Ultra-Low Bits
Amir Reza Mirzaei
Yuqiao Wen
Yanshuai Cao
Lili Mou
MQ
453
0
0
30 Oct 2025
MISA: Memory-Efficient LLMs Optimization with Module-wise Importance Sampling
MISA: Memory-Efficient LLMs Optimization with Module-wise Importance Sampling
Yuxi Liu
Renjia Deng
Yutong He
Xue Wang
Tao Yao
Kun Yuan
110
0
0
28 Oct 2025
Conditions for Catastrophic Forgetting in Multilingual Translation
Conditions for Catastrophic Forgetting in Multilingual Translation
Danni Liu
Jan Niehues
CLL
152
0
0
22 Oct 2025
Continual Learning via Sparse Memory Finetuning
Continual Learning via Sparse Memory Finetuning
Jessy Lin
Luke Zettlemoyer
Gargi Ghosh
Wen-tau Yih
Aram H. Markosyan
Vincent-Pierre Berges
Barlas Oğuz
KELMCLL
132
0
0
16 Oct 2025
OPLoRA: Orthogonal Projection LoRA Prevents Catastrophic Forgetting during Parameter-Efficient Fine-Tuning
OPLoRA: Orthogonal Projection LoRA Prevents Catastrophic Forgetting during Parameter-Efficient Fine-Tuning
Yifeng Xiong
Xiaohui Xie
CLL
464
2
0
14 Oct 2025
Enhancing Large Language Model Reasoning via Selective Critical Token Fine-Tuning
Enhancing Large Language Model Reasoning via Selective Critical Token Fine-Tuning
Zhiwen Ruan
Yixia Li
He Zhu
Yun Chen
P. Li
Yang Liu
Guanhua Chen
LRM
110
1
0
13 Oct 2025
CoLoR-GAN: Continual Few-Shot Learning with Low-Rank Adaptation in Generative Adversarial Networks
CoLoR-GAN: Continual Few-Shot Learning with Low-Rank Adaptation in Generative Adversarial Networks
Munsif Ali
L. Rossi
Massimo Bertozzi
CLLGAN
200
1
0
13 Oct 2025
Scaling Language-Centric Omnimodal Representation Learning
Scaling Language-Centric Omnimodal Representation Learning
Chenghao Xiao
Hou Pong Chan
Hao Zhang
Weiwen Xu
Mahani Aljunied
Yu Rong
120
0
0
13 Oct 2025
CTR-LoRA: Curvature-Aware and Trust-Region Guided Low-Rank Adaptation for Large Language Models
CTR-LoRA: Curvature-Aware and Trust-Region Guided Low-Rank Adaptation for Large Language Models
Zhuxuanzi Wang
Mingqiao Mo
Xi Xiao
Chen Liu
Chenrui Ma
Yunbei Zhang
Xiao Wang
Smita Krishnaswamy
Tianyang Wang
111
0
0
11 Oct 2025
Maximum In-Support Return Modeling for Dynamic Recommendation with Language Model Prior
Maximum In-Support Return Modeling for Dynamic Recommendation with Language Model Prior
Xiaocong Chen
Siyu Wang
Lina Yao
OffRLAI4TS
76
0
0
09 Oct 2025
How to Teach Large Multimodal Models New Skills
How to Teach Large Multimodal Models New Skills
Zhen Zhu
Yiming Gong
Yao Xiao
Yaoyao Liu
Derek Hoiem
MLLMCLLKELM
165
0
0
09 Oct 2025
Teamwork: Collaborative Diffusion with Low-rank Coordination and Adaptation
Teamwork: Collaborative Diffusion with Low-rank Coordination and Adaptation
Sam Sartor
Pieter Peers
DiffM
132
1
0
07 Oct 2025
MASA: Rethinking the Representational Bottleneck in LoRA with Multi-A Shared Adaptation
MASA: Rethinking the Representational Bottleneck in LoRA with Multi-A Shared Adaptation
Qin Dong
Yuntian Tang
Heming Jia
Yunhang Shen
Bohan Jia
Wenxuan Huang
Lianyue Zhang
Jiao Xie
Shaohui Lin
MoE
76
0
0
07 Oct 2025
Optimizing Fine-Tuning through Advanced Initialization Strategies for Low-Rank Adaptation
Optimizing Fine-Tuning through Advanced Initialization Strategies for Low-Rank Adaptation
Yongfu Xue
AI4CE
120
0
0
04 Oct 2025
Skip-It? Theoretical Conditions for Layer Skipping in Vision-Language Models
Skip-It? Theoretical Conditions for Layer Skipping in Vision-Language Models
Max Hartman
Vidhata Arjun Jayaraman
Moulik Choraria
Akhil Bhimaraju
Lav Varshney
VLM
364
0
0
29 Sep 2025
Effective Quantization of Muon Optimizer States
Effective Quantization of Muon Optimizer States
Aman Gupta
Rafael Celente
Abhishek Shivanna
D. T. Braithwaite
Gregory Dexter
Shao Tang
Hiroto Udagawa
Daniel Silva
R. Ramanath
S. Keerthi
MQ
128
0
0
27 Sep 2025
MolSpectLLM: A Molecular Foundation Model Bridging Spectroscopy, Molecule Elucidation, and 3D Structure Generation
MolSpectLLM: A Molecular Foundation Model Bridging Spectroscopy, Molecule Elucidation, and 3D Structure Generation
Shuaike Shen
Jiaqing Xie
Zhuo Yang
Antong Zhang
S. Sun
Ben Gao
Tianfan Fu
Biqing Qi
Yuqiang Li
100
0
0
26 Sep 2025
Unsupervised Defect Detection for Surgical Instruments
Unsupervised Defect Detection for Surgical Instruments
J. Huang
Yichi Zhang
Jingxi Yu
Wei Chen
Seunghyun Hwang
Qiang Qiu
A. Reibman
Edward J. Delp
Fengqing Zhu
68
0
0
25 Sep 2025
SFT Doesn't Always Hurt General Capabilities: Revisiting Domain-Specific Fine-Tuning in LLMs
SFT Doesn't Always Hurt General Capabilities: Revisiting Domain-Specific Fine-Tuning in LLMs
J. Lin
Zhongruo Wang
Kun Qian
Tian Wang
Arvind Srinivasan
...
Weiqi Zhang
Sujay Sanghavi
C. L. P. Chen
Hyokun Yun
Lihong Li
CLL
314
1
0
25 Sep 2025
CR-Net: Scaling Parameter-Efficient Training with Cross-Layer Low-Rank Structure
CR-Net: Scaling Parameter-Efficient Training with Cross-Layer Low-Rank Structure
Boao Kong
Junzhu Liang
Yuxi Liu
Renjia Deng
Kun Yuan
155
1
0
23 Sep 2025
Latent Traits and Cross-Task Transfer: Deconstructing Dataset Interactions in LLM Fine-tuning
Latent Traits and Cross-Task Transfer: Deconstructing Dataset Interactions in LLM Fine-tuningJoint Conference on Lexical and Computational Semantics (*SEM), 2025
Shambhavi Krishna
Atharva Naik
Chaitali Agarwal
Sudharshan Govindan
Taesung Lee
Haw-Shiuan Chang
119
0
0
17 Sep 2025
Conan-Embedding-v2: Training an LLM from Scratch for Text Embeddings
Conan-Embedding-v2: Training an LLM from Scratch for Text Embeddings
Shiyu Li
Yang Tang
Ruijie Liu
Shi-Zhe Chen
Xi Chen
73
1
0
16 Sep 2025
Continually Adding New Languages to Multilingual Language Models
Continually Adding New Languages to Multilingual Language Models
A. Owodunni
Sachin Kumar
CLLKELMMoMe
173
2
0
14 Sep 2025
Domain Adaptation of LLMs for Process Data
Domain Adaptation of LLMs for Process Data
Rafael Oyamada
Jari Peeperkorn
Jochen De Weerdt
Johannes De Smedt
132
1
0
03 Sep 2025
Incident Analysis for AI Agents
Incident Analysis for AI Agents
Carson Ezell
Xavier Roberts-Gaal
Alan Chan
160
2
0
19 Aug 2025
LoRAtorio: An intrinsic approach to LoRA Skill Composition
LoRAtorio: An intrinsic approach to LoRA Skill Composition
Niki Foteinopoulou
Ignas Budvytis
Stephan Liwicki
MoMe
125
0
0
15 Aug 2025
KL-based self-distillation for large language models
KL-based self-distillation for large language models
Max Rehman Linder
76
0
0
14 Aug 2025
Resource-Limited Joint Multimodal Sentiment Reasoning and Classification via Chain-of-Thought Enhancement and Distillation
Resource-Limited Joint Multimodal Sentiment Reasoning and Classification via Chain-of-Thought Enhancement and Distillation
Haonan Shangguan
Xiaocui Yang
Shi Feng
Daling Wang
Yifei Zhang
Ge Yu
LRM
104
0
0
07 Aug 2025
Unveiling Over-Memorization in Finetuning LLMs for Reasoning Tasks
Unveiling Over-Memorization in Finetuning LLMs for Reasoning Tasks
Zhiwen Ruan
Yun-Nung Chen
Yutao Hou
Peng Li
Yang Liu
Guanhua Chen
160
1
0
06 Aug 2025
Continual Learning for VLMs: A Survey and Taxonomy Beyond Forgetting
Continual Learning for VLMs: A Survey and Taxonomy Beyond Forgetting
Yuyang Liu
Qiuhe Hong
Linlan Huang
Alexandra Gomez-Villa
Dipam Goswami
Xialei Liu
Joost van de Weijer
Yonghong Tian
CLLKELMVLM
189
0
0
06 Aug 2025
GeRe: Towards Efficient Anti-Forgetting in Continual Learning of LLM via General Samples Replay
GeRe: Towards Efficient Anti-Forgetting in Continual Learning of LLM via General Samples Replay
Yunan Zhang
Shuoran Jiang
Mengchen Zhao
Yuefeng Li
Yang Fan
Xiangping Wu
Qingcai Chen
KELMCLL
115
1
0
06 Aug 2025
Trustworthy Reasoning: Evaluating and Enhancing Factual Accuracy in LLM Intermediate Thought Processes
Trustworthy Reasoning: Evaluating and Enhancing Factual Accuracy in LLM Intermediate Thought Processes
Rui Jiao
Yue Zhang
Jinku Li
LRM
165
0
0
25 Jul 2025
BiLO: Bilevel Local Operator Learning for PDE Inverse Problems. Part II: Efficient Uncertainty Quantification with Low-Rank Adaptation
BiLO: Bilevel Local Operator Learning for PDE Inverse Problems. Part II: Efficient Uncertainty Quantification with Low-Rank Adaptation
Ray Zirui Zhang
Christopher E. Miles
Xiaohui Xie
John S. Lowengrub
150
0
0
22 Jul 2025
PUSA V1.0: Surpassing Wan-I2V with $500 Training Cost by Vectorized Timestep Adaptation
PUSA V1.0: Surpassing Wan-I2V with $500 Training Cost by Vectorized Timestep Adaptation
Yaofang Liu
Y. Ren
Aitor Artola
Yuxuan Hu
Xiaodong Cun
...
Raymond H. F. Chan
Suiyun Zhang
Rui Liu
Dandan Tu
Jean-Michel Morel
DiffMVGen
159
1
0
22 Jul 2025
A Practical Investigation of Spatially-Controlled Image Generation with Transformers
A Practical Investigation of Spatially-Controlled Image Generation with Transformers
Guoxuan Xia
Harleen Hanspal
Petru-Daniel Tudosiu
Shifeng Zhang
Sarah Parisot
191
0
0
21 Jul 2025
LoRA meets Riemannion: Muon Optimizer for Parametrization-independent Low-Rank Adapters
LoRA meets Riemannion: Muon Optimizer for Parametrization-independent Low-Rank Adapters
Vladimir Bogachev
Vladimir Aletov
Alexander Molozhavenko
Denis Bobkov
Vera Soboleva
Aibek Alanov
Maxim Rakhuba
125
1
0
16 Jul 2025
Small Batch Size Training for Language Models: When Vanilla SGD Works, and Why Gradient Accumulation Is Wasteful
Small Batch Size Training for Language Models: When Vanilla SGD Works, and Why Gradient Accumulation Is Wasteful
Martin Marek
Sanae Lotfi
Aditya Somasundaram
A. Wilson
Micah Goldblum
LRM
352
11
0
09 Jul 2025
LoSiA: Efficient High-Rank Fine-Tuning via Subnet Localization and Optimization
LoSiA: Efficient High-Rank Fine-Tuning via Subnet Localization and Optimization
Xujia Wang
Yunjia Qi
Bin Xu
176
0
0
06 Jul 2025
Pay Attention to Small Weights
Pay Attention to Small Weights
Chao Zhou
Tom Jacobs
Advait Gadhikar
R. Burkholz
142
0
0
26 Jun 2025
Little By Little: Continual Learning via Self-Activated Sparse Mixture-of-Rank Adaptive Learning
Little By Little: Continual Learning via Self-Activated Sparse Mixture-of-Rank Adaptive Learning
Haodong Lu
Chongyang Zhao
Jason Xue
Lina Yao
Kristen Moore
Dong Gong
CLLMoMeMoE
214
2
0
26 Jun 2025
ReCode: Updating Code API Knowledge with Reinforcement Learning
ReCode: Updating Code API Knowledge with Reinforcement Learning
Haoze Wu
Yunzhi Yao
Wenhao Yu
Ningyu Zhang
SyDa
259
4
0
25 Jun 2025
Revisiting LoRA through the Lens of Parameter Redundancy: Spectral Encoding Helps
Revisiting LoRA through the Lens of Parameter Redundancy: Spectral Encoding HelpsAnnual Meeting of the Association for Computational Linguistics (ACL), 2025
Jiashun Cheng
Chenyi Zi
Polydoros Giannouris
Ziqi Gao
Yuhan Li
Jia Li
Fugee Tsung
196
0
0
20 Jun 2025
Improving LoRA with Variational Learning
Improving LoRA with Variational Learning
Bai Cong
Nico Daheim
Yuesong Shen
Rio Yokota
Mohammad Emtiyaz Khan
Thomas Möllenhoff
199
1
0
17 Jun 2025
Massive Supervised Fine-tuning Experiments Reveal How Data, Layer, and Training Factors Shape LLM Alignment Quality
Massive Supervised Fine-tuning Experiments Reveal How Data, Layer, and Training Factors Shape LLM Alignment Quality
Yuto Harada
Yusuke Yamauchi
Yusuke Oda
Yohei Oseki
Yusuke Miyao
Yu Takagi
ALM
188
3
0
17 Jun 2025
LARGO: Low-Rank Regulated Gradient Projection for Robust Parameter Efficient Fine-Tuning
LARGO: Low-Rank Regulated Gradient Projection for Robust Parameter Efficient Fine-Tuning
Haotian Zhang
Liu Liu
Baosheng Yu
Jiayan Qiu
Yanwei Ren
Xianglong Liu
182
0
0
14 Jun 2025
SOFT: Selective Data Obfuscation for Protecting LLM Fine-tuning against Membership Inference Attacks
SOFT: Selective Data Obfuscation for Protecting LLM Fine-tuning against Membership Inference Attacks
Kaiyuan Zhang
Siyuan Cheng
Hanxi Guo
Yuetian Chen
Zian Su
...
Yuntao Du
Charles Fleming
Jayanth Srinivasa
Xiangyu Zhang
Ninghui Li
AAML
356
5
0
12 Jun 2025
123
Next