ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1811.01088
  4. Cited By
Sentence Encoders on STILTs: Supplementary Training on Intermediate
  Labeled-data Tasks

Sentence Encoders on STILTs: Supplementary Training on Intermediate Labeled-data Tasks

2 November 2018
Jason Phang
Thibault Févry
Samuel R. Bowman
ArXivPDFHTML

Papers citing "Sentence Encoders on STILTs: Supplementary Training on Intermediate Labeled-data Tasks"

28 / 78 papers shown
Title
Entailment as Few-Shot Learner
Entailment as Few-Shot Learner
Sinong Wang
Han Fang
Madian Khabsa
Hanzi Mao
Hao Ma
30
183
0
29 Apr 2021
AMMU : A Survey of Transformer-based Biomedical Pretrained Language
  Models
AMMU : A Survey of Transformer-based Biomedical Pretrained Language Models
Katikapalli Subramanyam Kalyan
A. Rajasekharan
S. Sangeetha
LM&MA
MedIm
18
163
0
16 Apr 2021
How to Train BERT with an Academic Budget
How to Train BERT with an Academic Budget
Peter Izsak
Moshe Berchansky
Omer Levy
10
111
0
15 Apr 2021
FEWS: Large-Scale, Low-Shot Word Sense Disambiguation with the
  Dictionary
FEWS: Large-Scale, Low-Shot Word Sense Disambiguation with the Dictionary
Terra Blevins
Mandar Joshi
Luke Zettlemoyer
22
19
0
16 Feb 2021
Making Pre-trained Language Models Better Few-shot Learners
Making Pre-trained Language Models Better Few-shot Learners
Tianyu Gao
Adam Fisch
Danqi Chen
241
1,916
0
31 Dec 2020
Pretrained Transformers for Text Ranking: BERT and Beyond
Pretrained Transformers for Text Ranking: BERT and Beyond
Jimmy J. Lin
Rodrigo Nogueira
Andrew Yates
VLM
219
608
0
13 Oct 2020
Conditionally Adaptive Multi-Task Learning: Improving Transfer Learning
  in NLP Using Fewer Parameters & Less Data
Conditionally Adaptive Multi-Task Learning: Improving Transfer Learning in NLP Using Fewer Parameters & Less Data
Jonathan Pilault
Amine Elhattami
C. Pal
CLL
MoE
19
89
0
19 Sep 2020
Self-Supervised Meta-Learning for Few-Shot Natural Language
  Classification Tasks
Self-Supervised Meta-Learning for Few-Shot Natural Language Classification Tasks
Trapit Bansal
Rishikesh Jha
Tsendsuren Munkhdalai
Andrew McCallum
SSL
VLM
20
87
0
17 Sep 2020
Critical Thinking for Language Models
Critical Thinking for Language Models
Gregor Betz
Christian Voigt
Kyle Richardson
SyDa
ReLM
LRM
AI4CE
18
35
0
15 Sep 2020
Transferability of Natural Language Inference to Biomedical Question
  Answering
Transferability of Natural Language Inference to Biomedical Question Answering
Minbyul Jeong
Mujeen Sung
Gangwoo Kim
Donghyeon Kim
Wonjin Yoon
J. Yoo
Jaewoo Kang
19
37
0
01 Jul 2020
Revisiting Few-sample BERT Fine-tuning
Revisiting Few-sample BERT Fine-tuning
Tianyi Zhang
Felix Wu
Arzoo Katiyar
Kilian Q. Weinberger
Yoav Artzi
30
441
0
10 Jun 2020
Language Models are Few-Shot Learners
Language Models are Few-Shot Learners
Tom B. Brown
Benjamin Mann
Nick Ryder
Melanie Subbiah
Jared Kaplan
...
Christopher Berner
Sam McCandlish
Alec Radford
Ilya Sutskever
Dario Amodei
BDL
15
39,839
0
28 May 2020
Pre-training Is (Almost) All You Need: An Application to Commonsense
  Reasoning
Pre-training Is (Almost) All You Need: An Application to Commonsense Reasoning
Alexandre Tamborrino
Nicola Pellicanò
B. Pannier
Pascal Voitot
Louise Naudin
LRM
6
62
0
29 Apr 2020
Train No Evil: Selective Masking for Task-Guided Pre-Training
Train No Evil: Selective Masking for Task-Guided Pre-Training
Yuxian Gu
Zhengyan Zhang
Xiaozhi Wang
Zhiyuan Liu
Maosong Sun
19
59
0
21 Apr 2020
Beyond Fine-tuning: Few-Sample Sentence Embedding Transfer
Beyond Fine-tuning: Few-Sample Sentence Embedding Transfer
Siddhant Garg
Rohit Kumar Sharma
Yingyu Liang
20
4
0
10 Apr 2020
Pre-trained Models for Natural Language Processing: A Survey
Pre-trained Models for Natural Language Processing: A Survey
Xipeng Qiu
Tianxiang Sun
Yige Xu
Yunfan Shao
Ning Dai
Xuanjing Huang
LM&MA
VLM
241
1,450
0
18 Mar 2020
HUBERT Untangles BERT to Improve Transfer across NLP Tasks
HUBERT Untangles BERT to Improve Transfer across NLP Tasks
M. Moradshahi
Hamid Palangi
M. Lam
P. Smolensky
Jianfeng Gao
21
16
0
25 Oct 2019
Exploring the Limits of Transfer Learning with a Unified Text-to-Text
  Transformer
Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer
Colin Raffel
Noam M. Shazeer
Adam Roberts
Katherine Lee
Sharan Narang
Michael Matena
Yanqi Zhou
Wei Li
Peter J. Liu
AIMat
59
19,391
0
23 Oct 2019
Mixout: Effective Regularization to Finetune Large-scale Pretrained
  Language Models
Mixout: Effective Regularization to Finetune Large-scale Pretrained Language Models
Cheolhyoung Lee
Kyunghyun Cho
Wanmo Kang
MoE
235
205
0
25 Sep 2019
Task Selection Policies for Multitask Learning
Task Selection Policies for Multitask Learning
John Glover
Chris Hokamp
OffRL
24
7
0
14 Jul 2019
BAM! Born-Again Multi-Task Networks for Natural Language Understanding
BAM! Born-Again Multi-Task Networks for Natural Language Understanding
Kevin Clark
Minh-Thang Luong
Urvashi Khandelwal
Christopher D. Manning
Quoc V. Le
19
228
0
10 Jul 2019
Transfer Learning for Risk Classification of Social Media Posts: Model
  Evaluation Study
Transfer Learning for Risk Classification of Social Media Posts: Model Evaluation Study
Derek Howard
M. Maslej
Justin Lee
Jacob Ritchie
G. Woollard
L. French
AI4MH
18
30
0
04 Jul 2019
Human vs. Muppet: A Conservative Estimate of Human Performance on the
  GLUE Benchmark
Human vs. Muppet: A Conservative Estimate of Human Performance on the GLUE Benchmark
Nikita Nangia
Samuel R. Bowman
ELM
ALM
20
75
0
24 May 2019
BoolQ: Exploring the Surprising Difficulty of Natural Yes/No Questions
BoolQ: Exploring the Surprising Difficulty of Natural Yes/No Questions
Christopher Clark
Kenton Lee
Ming-Wei Chang
Tom Kwiatkowski
Michael Collins
Kristina Toutanova
17
1,386
0
24 May 2019
Story Ending Prediction by Transferable BERT
Story Ending Prediction by Transferable BERT
Zhongyang Li
Xiao Ding
Ting Liu
18
52
0
17 May 2019
Linguistic Analysis of Pretrained Sentence Encoders with Acceptability
  Judgments
Linguistic Analysis of Pretrained Sentence Encoders with Acceptability Judgments
Alex Warstadt
Samuel R. Bowman
11
23
0
11 Jan 2019
What you can cram into a single vector: Probing sentence embeddings for
  linguistic properties
What you can cram into a single vector: Probing sentence embeddings for linguistic properties
Alexis Conneau
Germán Kruszewski
Guillaume Lample
Loïc Barrault
Marco Baroni
199
882
0
03 May 2018
GLUE: A Multi-Task Benchmark and Analysis Platform for Natural Language
  Understanding
GLUE: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding
Alex Jinpeng Wang
Amanpreet Singh
Julian Michael
Felix Hill
Omer Levy
Samuel R. Bowman
ELM
294
6,943
0
20 Apr 2018
Previous
12