ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2026 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2004.10964
  4. Cited By
Don't Stop Pretraining: Adapt Language Models to Domains and Tasks
v1v2v3 (latest)

Don't Stop Pretraining: Adapt Language Models to Domains and Tasks

23 April 2020
Suchin Gururangan
Ana Marasović
Swabha Swayamdipta
Kyle Lo
Iz Beltagy
Doug Downey
Noah A. Smith
    VLMAI4CECLL
ArXiv (abs)PDFHTML

Papers citing "Don't Stop Pretraining: Adapt Language Models to Domains and Tasks"

19 / 1,369 papers shown
Investigating Pretrained Language Models for Graph-to-Text Generation
Investigating Pretrained Language Models for Graph-to-Text Generation
Leonardo F. R. Ribeiro
Martin Schmitt
Hinrich Schütze
Iryna Gurevych
353
226
0
16 Jul 2020
Composed Fine-Tuning: Freezing Pre-Trained Denoising Autoencoders for
  Improved Generalization
Composed Fine-Tuning: Freezing Pre-Trained Denoising Autoencoders for Improved Generalization
Sang Michael Xie
Tengyu Ma
Abigail Z. Jacobs
354
17
0
29 Jun 2020
Modeling Subjective Assessments of Guilt in Newspaper Crime Narratives
Modeling Subjective Assessments of Guilt in Newspaper Crime Narratives
Elisa Kreiss
Zijian Wang
Christopher Potts
135
1
0
17 Jun 2020
Foreseeing the Benefits of Incidental Supervision
Foreseeing the Benefits of Incidental Supervision
Hangfeng He
Mingyuan Zhang
Qiang Ning
Dan Roth
221
11
0
09 Jun 2020
Neural Unsupervised Domain Adaptation in NLP---A Survey
Neural Unsupervised Domain Adaptation in NLP---A SurveyInternational Conference on Computational Linguistics (COLING), 2020
Alan Ramponi
Barbara Plank
OOD
370
278
0
31 May 2020
CausaLM: Causal Model Explanation Through Counterfactual Language Models
CausaLM: Causal Model Explanation Through Counterfactual Language ModelsInternational Conference on Computational Logic (ICCL), 2020
Amir Feder
Nadav Oved
Uri Shalit
Roi Reichart
CMLLRM
558
179
0
27 May 2020
BERTweet: A pre-trained language model for English Tweets
BERTweet: A pre-trained language model for English Tweets
Dat Quoc Nguyen
Thanh Tien Vu
A. Nguyen
VLM
275
1,010
0
20 May 2020
Pre-training technique to localize medical BERT and enhance biomedical
  BERT
Pre-training technique to localize medical BERT and enhance biomedical BERT
Shoya Wada
Toshihiro Takeda
S. Manabe
Shozo Konishi
Jun Kamohara
Y. Matsumura
LM&MA
243
18
0
14 May 2020
Evidence Inference 2.0: More Data, Better Models
Evidence Inference 2.0: More Data, Better Models
Jay DeYoung
Eric P. Lehman
Benjamin E. Nye
Iain J. Marshall
Byron C. Wallace
309
74
0
08 May 2020
Fact or Fiction: Verifying Scientific Claims
Fact or Fiction: Verifying Scientific ClaimsConference on Empirical Methods in Natural Language Processing (EMNLP), 2020
Aman Rangapur
Shanchuan Lin
Kyle Lo
Lucy Lu Wang
Madeleine van Zuylen
Arman Cohan
Hannaneh Hajishirzi
HAI
569
593
0
30 Apr 2020
DomBERT: Domain-oriented Language Model for Aspect-based Sentiment
  Analysis
DomBERT: Domain-oriented Language Model for Aspect-based Sentiment AnalysisFindings (Findings), 2020
Hu Xu
Bing-Quan Liu
Lei Shu
Philip S. Yu
163
56
0
28 Apr 2020
Generative Data Augmentation for Commonsense Reasoning
Generative Data Augmentation for Commonsense Reasoning
Yiben Yang
Chaitanya Malaviya
Jared Fernandez
Swabha Swayamdipta
Ronan Le Bras
Ji-ping Wang
Chandra Bhagavatula
Yejin Choi
Doug Downey
LRM
241
89
0
24 Apr 2020
Privacy at Scale: Introducing the PrivaSeer Corpus of Web Privacy
  Policies
Privacy at Scale: Introducing the PrivaSeer Corpus of Web Privacy Policies
Mukund Srinath
Shomir Wilson
C. Lee Giles
PILM
201
62
0
23 Apr 2020
Train No Evil: Selective Masking for Task-Guided Pre-Training
Train No Evil: Selective Masking for Task-Guided Pre-TrainingConference on Empirical Methods in Natural Language Processing (EMNLP), 2020
Yuxian Gu
Zhengyan Zhang
Xiaozhi Wang
Zhiyuan Liu
Maosong Sun
257
62
0
21 Apr 2020
Unsupervised Domain Clusters in Pretrained Language Models
Unsupervised Domain Clusters in Pretrained Language ModelsAnnual Meeting of the Association for Computational Linguistics (ACL), 2020
Roee Aharoni
Yoav Goldberg
361
269
0
05 Apr 2020
Pre-trained Models for Natural Language Processing: A Survey
Pre-trained Models for Natural Language Processing: A SurveyScience China Technological Sciences (Sci China Technol Sci), 2020
Xipeng Qiu
Tianxiang Sun
Yige Xu
Yunfan Shao
Ning Dai
Xuanjing Huang
LM&MAVLM
1.1K
1,623
0
18 Mar 2020
LAMBERT: Layout-Aware (Language) Modeling for information extraction
LAMBERT: Layout-Aware (Language) Modeling for information extractionIEEE International Conference on Document Analysis and Recognition (ICDAR), 2020
Lukasz Garncarek
Rafal Powalski
Tomasz Stanislawek
Bartosz Topolski
Piotr Halama
M. Turski
Filip Graliñski
335
94
0
19 Feb 2020
Explaining Relationships Between Scientific Documents
Explaining Relationships Between Scientific DocumentsAnnual Meeting of the Association for Computational Linguistics (ACL), 2020
Kelvin Luu
Xinyi Wu
Rik Koncel-Kedziorski
Kyle Lo
Isabel Cachola
Noah A. Smith
360
55
0
02 Feb 2020
Billion-scale similarity search with GPUs
Billion-scale similarity search with GPUsIEEE Transactions on Big Data (TBD), 2017
Jeff Johnson
Matthijs Douze
Edouard Grave
970
4,531
0
28 Feb 2017
Previous
123...262728