ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2026 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2106.10199
  4. Cited By
BitFit: Simple Parameter-efficient Fine-tuning for Transformer-based
  Masked Language-models
v1v2v3v4v5 (latest)

BitFit: Simple Parameter-efficient Fine-tuning for Transformer-based Masked Language-models

Annual Meeting of the Association for Computational Linguistics (ACL), 2021
18 June 2021
Elad Ben-Zaken
Shauli Ravfogel
Yoav Goldberg
ArXiv (abs)PDFHTML

Papers citing "BitFit: Simple Parameter-efficient Fine-tuning for Transformer-based Masked Language-models"

50 / 968 papers shown
Parameter-efficient is not sufficient: Exploring Parameter, Memory, and
  Time Efficient Adapter Tuning for Dense Predictions
Parameter-efficient is not sufficient: Exploring Parameter, Memory, and Time Efficient Adapter Tuning for Dense PredictionsACM Multimedia (ACM MM), 2023
Dongshuo Yin
Xueting Han
Bin Li
Hao Feng
Jinghua Bai
VPVLM
273
27
0
16 Jun 2023
One-for-All: Generalized LoRA for Parameter-Efficient Fine-tuning
One-for-All: Generalized LoRA for Parameter-Efficient Fine-tuning
Arnav Chavan
Zhuang Liu
D. K. Gupta
Eric P. Xing
Zhiqiang Shen
314
109
0
13 Jun 2023
Morphosyntactic probing of multilingual BERT models
Morphosyntactic probing of multilingual BERT modelsNatural Language Engineering (NLE), 2023
Judit Ács
Endre Hamerlik
Roy Schwartz
Noah A. Smith
András Kornai
187
17
0
09 Jun 2023
Git-Theta: A Git Extension for Collaborative Development of Machine
  Learning Models
Git-Theta: A Git Extension for Collaborative Development of Machine Learning ModelsInternational Conference on Machine Learning (ICML), 2023
Nikhil Kandpal
Brian Lester
Mohammed Muqeeth
Anisha Mascarenhas
Monty Evans
Vishal Baskaran
Tenghao Huang
Haokun Liu
Colin Raffel
VLM
202
15
0
07 Jun 2023
ViDA: Homeostatic Visual Domain Adapter for Continual Test Time
  Adaptation
ViDA: Homeostatic Visual Domain Adapter for Continual Test Time AdaptationInternational Conference on Learning Representations (ICLR), 2023
Jiaming Liu
Senqiao Yang
Peidong Jia
Renrui Zhang
Ming Lu
Yandong Guo
Wei Xue
Shanghang Zhang
TTAOODVLM
397
57
0
07 Jun 2023
Inference-Time Intervention: Eliciting Truthful Answers from a Language
  Model
Inference-Time Intervention: Eliciting Truthful Answers from a Language ModelNeural Information Processing Systems (NeurIPS), 2023
Kenneth Li
Oam Patel
Fernanda Viégas
Hanspeter Pfister
Martin Wattenberg
KELMHILM
731
826
0
06 Jun 2023
Sensitivity-Aware Finetuning for Accuracy Recovery on Deep Learning
  Hardware
Sensitivity-Aware Finetuning for Accuracy Recovery on Deep Learning Hardware
Lakshmi Nair
D. Bunandar
135
0
0
05 Jun 2023
Cross-Lingual Transfer with Target Language-Ready Task Adapters
Cross-Lingual Transfer with Target Language-Ready Task AdaptersAnnual Meeting of the Association for Computational Linguistics (ACL), 2023
Marinela Parović
Alan Ansell
Ivan Vulić
Anna Korhonen
179
14
0
05 Jun 2023
Prompt to be Consistent is Better than Self-Consistent? Few-Shot and
  Zero-Shot Fact Verification with Pre-trained Language Models
Prompt to be Consistent is Better than Self-Consistent? Few-Shot and Zero-Shot Fact Verification with Pre-trained Language ModelsAnnual Meeting of the Association for Computational Linguistics (ACL), 2023
Fengzhu Zeng
Wei Gao
179
10
0
05 Jun 2023
bgGLUE: A Bulgarian General Language Understanding Evaluation Benchmark
bgGLUE: A Bulgarian General Language Understanding Evaluation BenchmarkAnnual Meeting of the Association for Computational Linguistics (ACL), 2023
Momchil Hardalov
Pepa Atanasova
Todor Mihaylov
G. Angelova
K. Simov
P. Osenova
Ves Stoyanov
Ivan Koychev
Preslav Nakov
Dragomir R. Radev
ELMFedML
254
9
0
04 Jun 2023
Exploring the Impact of Model Scaling on Parameter-Efficient Tuning
Exploring the Impact of Model Scaling on Parameter-Efficient TuningConference on Empirical Methods in Natural Language Processing (EMNLP), 2023
Yusheng Su
Chi-Min Chan
Jiali Cheng
Yujia Qin
Yankai Lin
...
Ning Ding
Xingzhi Sun
Guotong Xie
Zhiyuan Liu
Maosong Sun
234
9
0
04 Jun 2023
Benchmarking Robustness of Adaptation Methods on Pre-trained
  Vision-Language Models
Benchmarking Robustness of Adaptation Methods on Pre-trained Vision-Language ModelsNeural Information Processing Systems (NeurIPS), 2023
Shuo Chen
Jindong Gu
Zhen Han
Yunpu Ma
Juil Sock
Volker Tresp
VPVLMVLM
351
31
0
03 Jun 2023
UniDiff: Advancing Vision-Language Models with Generative and
  Discriminative Learning
UniDiff: Advancing Vision-Language Models with Generative and Discriminative Learning
Xiao Dong
Runhu Huang
Xiaoyong Wei
Zequn Jie
Jianxing Yu
Jian Yin
Xiaodan Liang
VLMDiffM
117
2
0
01 Jun 2023
Jointly Reparametrized Multi-Layer Adaptation for Efficient and Private
  Tuning
Jointly Reparametrized Multi-Layer Adaptation for Efficient and Private TuningAnnual Meeting of the Association for Computational Linguistics (ACL), 2023
Umang Gupta
Aram Galstyan
Greg Ver Steeg
203
3
0
30 May 2023
PreQuant: A Task-agnostic Quantization Approach for Pre-trained Language
  Models
PreQuant: A Task-agnostic Quantization Approach for Pre-trained Language ModelsAnnual Meeting of the Association for Computational Linguistics (ACL), 2023
Zhuocheng Gong
Jiahao Liu
Qifan Wang
Yang Yang
Jingang Wang
Wei Wu
Yunsen Xian
Dongyan Zhao
Rui Yan
MQ
323
6
0
30 May 2023
Universality and Limitations of Prompt Tuning
Universality and Limitations of Prompt TuningNeural Information Processing Systems (NeurIPS), 2023
Yihan Wang
Jatin Chauhan
Wei Wang
Cho-Jui Hsieh
489
28
0
30 May 2023
Domain Specialization as the Key to Make Large Language Models
  Disruptive: A Comprehensive Survey
Domain Specialization as the Key to Make Large Language Models Disruptive: A Comprehensive SurveyACM Computing Surveys (ACM Comput. Surv.), 2023
Chen Ling
Xujiang Zhao
Jiaying Lu
Chengyuan Deng
Can Zheng
...
Chris White
Quanquan Gu
Jian Pei
Carl Yang
Bo Pan
ALM
401
211
0
30 May 2023
Efficient Storage of Fine-Tuned Models via Low-Rank Approximation of
  Weight Residuals
Efficient Storage of Fine-Tuned Models via Low-Rank Approximation of Weight Residuals
Simo Ryu
S. Seo
Jaejun Yoo
220
11
0
28 May 2023
Plug-and-Play Knowledge Injection for Pre-trained Language Models
Plug-and-Play Knowledge Injection for Pre-trained Language ModelsAnnual Meeting of the Association for Computational Linguistics (ACL), 2023
Zhengyan Zhang
Zhiyuan Zeng
Yankai Lin
Huadong Wang
Deming Ye
...
Xu Han
Zhiyuan Liu
Peng Li
Maosong Sun
Jie Zhou
KELM
228
16
0
28 May 2023
Stochastic Bridges as Effective Regularizers for Parameter-Efficient
  Tuning
Stochastic Bridges as Effective Regularizers for Parameter-Efficient TuningAnnual Meeting of the Association for Computational Linguistics (ACL), 2023
Weize Chen
Xu Han
Yankai Lin
Zhiyuan Liu
Maosong Sun
Jie Zhou
90
1
0
28 May 2023
Plug-and-Play Document Modules for Pre-trained Models
Plug-and-Play Document Modules for Pre-trained ModelsAnnual Meeting of the Association for Computational Linguistics (ACL), 2023
Chaojun Xiao
Zhengyan Zhang
Xu Han
Chi-Min Chan
Yankai Lin
Zhiyuan Liu
Xiangyang Li
Zhonghua Li
Bo Zhao
Maosong Sun
KELM
239
9
0
28 May 2023
Fine-tuning Happens in Tiny Subspaces: Exploring Intrinsic Task-specific
  Subspaces of Pre-trained Language Models
Fine-tuning Happens in Tiny Subspaces: Exploring Intrinsic Task-specific Subspaces of Pre-trained Language ModelsAnnual Meeting of the Association for Computational Linguistics (ACL), 2023
Zhong Zhang
Bang Liu
Junming Shao
214
16
0
27 May 2023
Weaker Than You Think: A Critical Look at Weakly Supervised Learning
Weaker Than You Think: A Critical Look at Weakly Supervised LearningAnnual Meeting of the Association for Computational Linguistics (ACL), 2023
D. Zhu
Xiaoyu Shen
Marius Mosbach
Andreas Stephan
Dietrich Klakow
NoLa
264
17
0
27 May 2023
Do We Really Need a Large Number of Visual Prompts?
Do We Really Need a Large Number of Visual Prompts?Neural Networks (Neural Netw.), 2023
Youngeun Kim
Yuhang Li
Abhishek Moitra
Ruokai Yin
Priyadarshini Panda
VLMVPVLM
241
11
0
26 May 2023
Few-shot Fine-tuning vs. In-context Learning: A Fair Comparison and
  Evaluation
Few-shot Fine-tuning vs. In-context Learning: A Fair Comparison and EvaluationAnnual Meeting of the Association for Computational Linguistics (ACL), 2023
Marius Mosbach
Tiago Pimentel
Haiqin Yang
Dietrich Klakow
Yanai Elazar
319
173
0
26 May 2023
Parameter-Efficient Fine-Tuning without Introducing New Latency
Parameter-Efficient Fine-Tuning without Introducing New LatencyAnnual Meeting of the Association for Computational Linguistics (ACL), 2023
Baohao Liao
Yan Meng
Christof Monz
133
68
0
26 May 2023
Neural Architecture Search for Parameter-Efficient Fine-tuning of Large
  Pre-trained Language Models
Neural Architecture Search for Parameter-Efficient Fine-tuning of Large Pre-trained Language ModelsAnnual Meeting of the Association for Computational Linguistics (ACL), 2023
Neal Lawton
Anoop Kumar
Govind Thattai
Aram Galstyan
Greg Ver Steeg
140
23
0
26 May 2023
READ: Recurrent Adaptation of Large Transformers
READ: Recurrent Adaptation of Large Transformers
Sida I. Wang
John Nguyen
Ke Li
Carole-Jean Wu
168
12
0
24 May 2023
MultiFusion: Fusing Pre-Trained Models for Multi-Lingual, Multi-Modal
  Image Generation
MultiFusion: Fusing Pre-Trained Models for Multi-Lingual, Multi-Modal Image GenerationNeural Information Processing Systems (NeurIPS), 2023
Marco Bellagente
Manuel Brack
H. Teufel
Felix Friedrich
Bjorn Deiseroth
...
Koen Oostermeijer
Andres Felipe Cruz Salinas
P. Schramowski
Kristian Kersting
Samuel Weinbach
369
28
0
24 May 2023
Winner-Take-All Column Row Sampling for Memory Efficient Adaptation of
  Language Model
Winner-Take-All Column Row Sampling for Memory Efficient Adaptation of Language ModelNeural Information Processing Systems (NeurIPS), 2023
Zirui Liu
Guanchu Wang
Shaochen Zhong
Zhaozhuo Xu
Daochen Zha
...
Zhimeng Jiang
Kaixiong Zhou
Vipin Chaudhary
Shuai Xu
Helen Zhou
263
21
0
24 May 2023
Towards Adaptive Prefix Tuning for Parameter-Efficient Language Model
  Fine-tuning
Towards Adaptive Prefix Tuning for Parameter-Efficient Language Model Fine-tuningAnnual Meeting of the Association for Computational Linguistics (ACL), 2023
Zhen-Ru Zhang
Chuanqi Tan
Haiyang Xu
Chengyu Wang
Yanjie Liang
Songfang Huang
134
52
0
24 May 2023
Bactrian-X: Multilingual Replicable Instruction-Following Models with
  Low-Rank Adaptation
Bactrian-X: Multilingual Replicable Instruction-Following Models with Low-Rank Adaptation
Jinyan Su
Fajri Koto
Minghao Wu
Alham Fikri Aji
Timothy Baldwin
ALM
212
85
0
24 May 2023
Few-shot Unified Question Answering: Tuning Models or Prompts?
Few-shot Unified Question Answering: Tuning Models or Prompts?Conference on Empirical Methods in Natural Language Processing (EMNLP), 2023
Srijan Bansal
Semih Yavuz
Bo Pang
Meghana Moorthy Bhat
Yingbo Zhou
204
2
0
23 May 2023
QLoRA: Efficient Finetuning of Quantized LLMs
QLoRA: Efficient Finetuning of Quantized LLMsNeural Information Processing Systems (NeurIPS), 2023
Tim Dettmers
Artidoro Pagnoni
Ari Holtzman
Luke Zettlemoyer
ALM
609
3,637
0
23 May 2023
TaskWeb: Selecting Better Source Tasks for Multi-task NLP
TaskWeb: Selecting Better Source Tasks for Multi-task NLPConference on Empirical Methods in Natural Language Processing (EMNLP), 2023
Joongwon Kim
Akari Asai
Gabriel Ilharco
Hannaneh Hajishirzi
253
13
0
22 May 2023
SPARSEFIT: Few-shot Prompting with Sparse Fine-tuning for Jointly
  Generating Predictions and Natural Language Explanations
SPARSEFIT: Few-shot Prompting with Sparse Fine-tuning for Jointly Generating Predictions and Natural Language ExplanationsAnnual Meeting of the Association for Computational Linguistics (ACL), 2023
Jesus Solano
Oana-Maria Camburu
Pasquale Minervini
220
4
0
22 May 2023
TADA: Efficient Task-Agnostic Domain Adaptation for Transformers
TADA: Efficient Task-Agnostic Domain Adaptation for TransformersAnnual Meeting of the Association for Computational Linguistics (ACL), 2023
Chia-Chien Hung
Lukas Lange
Jannik Strötgen
292
13
0
22 May 2023
Prefix Propagation: Parameter-Efficient Tuning for Long Sequences
Prefix Propagation: Parameter-Efficient Tuning for Long SequencesAnnual Meeting of the Association for Computational Linguistics (ACL), 2023
Jonathan Li
Will Aitken
R. Bhambhoria
Xiao-Dan Zhu
184
15
0
20 May 2023
ToolkenGPT: Augmenting Frozen Language Models with Massive Tools via
  Tool Embeddings
ToolkenGPT: Augmenting Frozen Language Models with Massive Tools via Tool EmbeddingsNeural Information Processing Systems (NeurIPS), 2023
Shibo Hao
Tianyang Liu
Zhen Wang
Zhiting Hu
RALMLLMAG
584
234
0
19 May 2023
A Parameter-Efficient Learning Approach to Arabic Dialect Identification
  with Pre-Trained General-Purpose Speech Model
A Parameter-Efficient Learning Approach to Arabic Dialect Identification with Pre-Trained General-Purpose Speech ModelInterspeech (Interspeech), 2023
S. Radhakrishnan
Chao-Han Huck Yang
S. Khan
N. Kiani
D. Gómez-Cabrero
Jesper N. Tegnér
145
23
0
18 May 2023
Ahead-of-Time P-Tuning
Ahead-of-Time P-Tuning
Daniil Gavrilov
Nikita Balagansky
102
1
0
18 May 2023
G-Adapter: Towards Structure-Aware Parameter-Efficient Transfer Learning
  for Graph Transformer Networks
G-Adapter: Towards Structure-Aware Parameter-Efficient Transfer Learning for Graph Transformer NetworksAAAI Conference on Artificial Intelligence (AAAI), 2023
Anchun Gui
Jinqiang Ye
Han Xiao
147
31
0
17 May 2023
When Gradient Descent Meets Derivative-Free Optimization: A Match Made
  in Black-Box Scenario
When Gradient Descent Meets Derivative-Free Optimization: A Match Made in Black-Box ScenarioAnnual Meeting of the Association for Computational Linguistics (ACL), 2023
Chengcheng Han
Liqing Cui
Renyu Zhu
Jiadong Wang
Polydoros Giannouris
Qiushi Sun
Xiang Li
Ming Gao
203
9
0
17 May 2023
Knowledge Card: Filling LLMs' Knowledge Gaps with Plug-in Specialized
  Language Models
Knowledge Card: Filling LLMs' Knowledge Gaps with Plug-in Specialized Language ModelsInternational Conference on Learning Representations (ICLR), 2023
Shangbin Feng
Weijia Shi
Yuyang Bai
Vidhisha Balachandran
Tianxing He
Yulia Tsvetkov
KELM
365
49
0
17 May 2023
Memorization for Good: Encryption with Autoregressive Language Models
Memorization for Good: Encryption with Autoregressive Language Models
Samuel Stevens
Yung-Chun Su
160
1
0
15 May 2023
Parameter-Efficient Fine-Tuning with Layer Pruning on Free-Text
  Sequence-to-Sequence Modeling
Parameter-Efficient Fine-Tuning with Layer Pruning on Free-Text Sequence-to-Sequence Modeling
Y. Zhu
Xuebing Yang
Yuanyuan Wu
Wensheng Zhang
MedIm
262
4
0
15 May 2023
Parameter-Efficient Fine-Tuning for Medical Image Analysis: The Missed
  Opportunity
Parameter-Efficient Fine-Tuning for Medical Image Analysis: The Missed OpportunityInternational Conference on Medical Imaging with Deep Learning (MIDL), 2023
Raman Dutt
Linus Ericsson
Pedro Sanchez
Sotirios A. Tsaftaris
Timothy M. Hospedales
MedIm
477
73
0
14 May 2023
Make Prompt-based Black-Box Tuning Colorful: Boosting Model
  Generalization from Three Orthogonal Perspectives
Make Prompt-based Black-Box Tuning Colorful: Boosting Model Generalization from Three Orthogonal PerspectivesInternational Conference on Language Resources and Evaluation (LREC), 2023
Qiushi Sun
Chengcheng Han
Polydoros Giannouris
Renyu Zhu
Jing Gong
Xiang Li
Ming Gao
VLM
108
9
0
14 May 2023
ArtGPT-4: Towards Artistic-understanding Large Vision-Language Models
  with Enhanced Adapter
ArtGPT-4: Towards Artistic-understanding Large Vision-Language Models with Enhanced Adapter
Zheng Yuan
HU Xue
Kun Wang
Yongming Liu
Kun Wang
VLMMLLM
372
12
0
12 May 2023
Incorporating Structured Representations into Pretrained Vision &
  Language Models Using Scene Graphs
Incorporating Structured Representations into Pretrained Vision & Language Models Using Scene GraphsConference on Empirical Methods in Natural Language Processing (EMNLP), 2023
Roei Herzig
Alon Mendelson
Leonid Karlinsky
Assaf Arbelle
Rogerio Feris
Trevor Darrell
Amir Globerson
VLM
295
39
0
10 May 2023
Previous
123...141516...181920
Next