ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1911.03437
  4. Cited By
SMART: Robust and Efficient Fine-Tuning for Pre-trained Natural Language
  Models through Principled Regularized Optimization

SMART: Robust and Efficient Fine-Tuning for Pre-trained Natural Language Models through Principled Regularized Optimization

8 November 2019
Haoming Jiang
Pengcheng He
Weizhu Chen
Xiaodong Liu
Jianfeng Gao
T. Zhao
ArXivPDFHTML

Papers citing "SMART: Robust and Efficient Fine-Tuning for Pre-trained Natural Language Models through Principled Regularized Optimization"

27 / 77 papers shown
Title
Taming Sparsely Activated Transformer with Stochastic Experts
Taming Sparsely Activated Transformer with Stochastic Experts
Simiao Zuo
Xiaodong Liu
Jian Jiao
Young Jin Kim
Hany Hassan
Ruofei Zhang
T. Zhao
Jianfeng Gao
MoE
37
108
0
08 Oct 2021
Using Psuedolabels for training Sentiment Classifiers makes the model
  generalize better across datasets
Using Psuedolabels for training Sentiment Classifiers makes the model generalize better across datasets
N. Reddy
Muktabh Mayank Srivastava
14
0
0
05 Oct 2021
Virtual Data Augmentation: A Robust and General Framework for
  Fine-tuning Pre-trained Models
Virtual Data Augmentation: A Robust and General Framework for Fine-tuning Pre-trained Models
Kun Zhou
Wayne Xin Zhao
Sirui Wang
Fuzheng Zhang
Wei Yu Wu
Ji-Rong Wen
AAML
13
7
0
13 Sep 2021
TREATED:Towards Universal Defense against Textual Adversarial Attacks
TREATED:Towards Universal Defense against Textual Adversarial Attacks
Bin Zhu
Zhaoquan Gu
Le Wang
Zhihong Tian
AAML
25
8
0
13 Sep 2021
Robust fine-tuning of zero-shot models
Robust fine-tuning of zero-shot models
Mitchell Wortsman
Gabriel Ilharco
Jong Wook Kim
Mike Li
Simon Kornblith
...
Raphael Gontijo-Lopes
Hannaneh Hajishirzi
Ali Farhadi
Hongseok Namkoong
Ludwig Schmidt
VLM
21
687
0
04 Sep 2021
How Does Adversarial Fine-Tuning Benefit BERT?
How Does Adversarial Fine-Tuning Benefit BERT?
J. Ebrahimi
Hao Yang
Wei Zhang
AAML
13
4
0
31 Aug 2021
Noise Stability Regularization for Improving BERT Fine-tuning
Noise Stability Regularization for Improving BERT Fine-tuning
Hang Hua
Xingjian Li
Dejing Dou
Chengzhong Xu
Jiebo Luo
11
42
0
10 Jul 2021
A Survey on Data Augmentation for Text Classification
A Survey on Data Augmentation for Text Classification
Markus Bayer
M. Kaufhold
Christian A. Reuter
23
332
0
07 Jul 2021
CLINE: Contrastive Learning with Semantic Negative Examples for Natural
  Language Understanding
CLINE: Contrastive Learning with Semantic Negative Examples for Natural Language Understanding
Dong Wang
Ning Ding
Pijian Li
Haitao Zheng
AAML
24
115
0
01 Jul 2021
Learn to Resolve Conversational Dependency: A Consistency Training
  Framework for Conversational Question Answering
Learn to Resolve Conversational Dependency: A Consistency Training Framework for Conversational Question Answering
Gangwoo Kim
Hyunjae Kim
Jungsoo Park
Jaewoo Kang
14
38
0
22 Jun 2021
An Empirical Survey of Data Augmentation for Limited Data Learning in
  NLP
An Empirical Survey of Data Augmentation for Limited Data Learning in NLP
Jiaao Chen
Derek Tam
Colin Raffel
Mohit Bansal
Diyi Yang
21
172
0
14 Jun 2021
Self-Guided Contrastive Learning for BERT Sentence Representations
Self-Guided Contrastive Learning for BERT Sentence Representations
Taeuk Kim
Kang Min Yoo
Sang-goo Lee
SSL
20
202
0
03 Jun 2021
Empirical Evaluation of Pre-trained Transformers for Human-Level NLP:
  The Role of Sample Size and Dimensionality
Empirical Evaluation of Pre-trained Transformers for Human-Level NLP: The Role of Sample Size and Dimensionality
Adithya V Ganesan
Matthew Matero
Aravind Reddy Ravula
Huy-Hien Vu
H. A. Schwartz
12
35
0
07 May 2021
Playing Lottery Tickets with Vision and Language
Playing Lottery Tickets with Vision and Language
Zhe Gan
Yen-Chun Chen
Linjie Li
Tianlong Chen
Yu Cheng
Shuohang Wang
Jingjing Liu
Lijuan Wang
Zicheng Liu
VLM
101
53
0
23 Apr 2021
Reweighting Augmented Samples by Minimizing the Maximal Expected Loss
Reweighting Augmented Samples by Minimizing the Maximal Expected Loss
Mingyang Yi
Lu Hou
Lifeng Shang
Xin Jiang
Qun Liu
Zhi-Ming Ma
8
19
0
16 Mar 2021
Contrastive Learning with Adversarial Perturbations for Conditional Text
  Generation
Contrastive Learning with Adversarial Perturbations for Conditional Text Generation
Seanie Lee
Dong Bok Lee
Sung Ju Hwang
13
105
0
14 Dec 2020
Calibrated Language Model Fine-Tuning for In- and Out-of-Distribution
  Data
Calibrated Language Model Fine-Tuning for In- and Out-of-Distribution Data
Lingkai Kong
Haoming Jiang
Yuchen Zhuang
Jie Lyu
T. Zhao
Chao Zhang
OODD
13
26
0
22 Oct 2020
Robust Optimization as Data Augmentation for Large-scale Graphs
Robust Optimization as Data Augmentation for Large-scale Graphs
Kezhi Kong
G. Li
Mucong Ding
Zuxuan Wu
Chen Zhu
Bernard Ghanem
Gavin Taylor
Tom Goldstein
87
74
0
19 Oct 2020
InfoBERT: Improving Robustness of Language Models from An Information
  Theoretic Perspective
InfoBERT: Improving Robustness of Language Models from An Information Theoretic Perspective
Boxin Wang
Shuohang Wang
Yu Cheng
Zhe Gan
R. Jia
Bo-wen Li
Jingjing Liu
AAML
36
113
0
05 Oct 2020
Conditionally Adaptive Multi-Task Learning: Improving Transfer Learning
  in NLP Using Fewer Parameters & Less Data
Conditionally Adaptive Multi-Task Learning: Improving Transfer Learning in NLP Using Fewer Parameters & Less Data
Jonathan Pilault
Amine Elhattami
C. Pal
CLL
MoE
19
89
0
19 Sep 2020
FarsTail: A Persian Natural Language Inference Dataset
FarsTail: A Persian Natural Language Inference Dataset
Hossein Amirkhani
Mohammad AzariJafari
Zohreh Pourjafari
Soroush Faridan-Jahromi
Zeinab Kouhkan
Azadeh Amirak
19
29
0
18 Sep 2020
STEAM: Self-Supervised Taxonomy Expansion with Mini-Paths
STEAM: Self-Supervised Taxonomy Expansion with Mini-Paths
Yue Yu
Yinghao Li
Jiaming Shen
Haoyang Feng
Jimeng Sun
Chao Zhang
18
58
0
18 Jun 2020
Large-Scale Adversarial Training for Vision-and-Language Representation
  Learning
Large-Scale Adversarial Training for Vision-and-Language Representation Learning
Zhe Gan
Yen-Chun Chen
Linjie Li
Chen Zhu
Yu Cheng
Jingjing Liu
ObjD
VLM
24
487
0
11 Jun 2020
DeBERTa: Decoding-enhanced BERT with Disentangled Attention
DeBERTa: Decoding-enhanced BERT with Disentangled Attention
Pengcheng He
Xiaodong Liu
Jianfeng Gao
Weizhu Chen
AAML
62
2,614
0
05 Jun 2020
FreeLB: Enhanced Adversarial Training for Natural Language Understanding
FreeLB: Enhanced Adversarial Training for Natural Language Understanding
Chen Zhu
Yu Cheng
Zhe Gan
S. Sun
Tom Goldstein
Jingjing Liu
AAML
221
436
0
25 Sep 2019
GLUE: A Multi-Task Benchmark and Analysis Platform for Natural Language
  Understanding
GLUE: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding
Alex Jinpeng Wang
Amanpreet Singh
Julian Michael
Felix Hill
Omer Levy
Samuel R. Bowman
ELM
294
6,943
0
20 Apr 2018
Mean teachers are better role models: Weight-averaged consistency
  targets improve semi-supervised deep learning results
Mean teachers are better role models: Weight-averaged consistency targets improve semi-supervised deep learning results
Antti Tarvainen
Harri Valpola
OOD
MoMe
244
1,275
0
06 Mar 2017
Previous
12