ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1906.08237
  4. Cited By
XLNet: Generalized Autoregressive Pretraining for Language Understanding

XLNet: Generalized Autoregressive Pretraining for Language Understanding

19 June 2019
Zhilin Yang
Zihang Dai
Yiming Yang
J. Carbonell
Ruslan Salakhutdinov
Quoc V. Le
    AI4CE
ArXivPDFHTML

Papers citing "XLNet: Generalized Autoregressive Pretraining for Language Understanding"

50 / 1,081 papers shown
Title
Quantitative Argument Summarization and Beyond: Cross-Domain Key Point
  Analysis
Quantitative Argument Summarization and Beyond: Cross-Domain Key Point Analysis
Roy Bar-Haim
Yoav Kantor
Lilach Eden
Roni Friedman
Dan Lahav
Noam Slonim
29
43
0
11 Oct 2020
SMYRF: Efficient Attention using Asymmetric Clustering
SMYRF: Efficient Attention using Asymmetric Clustering
Giannis Daras
Nikita Kitaev
Augustus Odena
A. Dimakis
23
44
0
11 Oct 2020
SJTU-NICT's Supervised and Unsupervised Neural Machine Translation
  Systems for the WMT20 News Translation Task
SJTU-NICT's Supervised and Unsupervised Neural Machine Translation Systems for the WMT20 News Translation Task
Z. Li
Hai Zhao
Rui Wang
Kehai Chen
Masao Utiyama
Eiichiro Sumita
29
15
0
11 Oct 2020
Automated Concatenation of Embeddings for Structured Prediction
Automated Concatenation of Embeddings for Structured Prediction
Xinyu Wang
Yong-jia Jiang
Nguyen Bach
Tao Wang
Zhongqiang Huang
Fei Huang
Kewei Tu
35
172
0
10 Oct 2020
Infusing Disease Knowledge into BERT for Health Question Answering,
  Medical Inference and Disease Name Recognition
Infusing Disease Knowledge into BERT for Health Question Answering, Medical Inference and Disease Name Recognition
Yun He
Ziwei Zhu
Yin Zhang
Qin Chen
James Caverlee
AI4MH
28
108
0
08 Oct 2020
What Can We Learn from Collective Human Opinions on Natural Language
  Inference Data?
What Can We Learn from Collective Human Opinions on Natural Language Inference Data?
Yixin Nie
Xiang Zhou
Mohit Bansal
6
128
0
07 Oct 2020
Poison Attacks against Text Datasets with Conditional Adversarially
  Regularized Autoencoder
Poison Attacks against Text Datasets with Conditional Adversarially Regularized Autoencoder
Alvin Chan
Yi Tay
Yew-Soon Ong
Aston Zhang
SILM
13
57
0
06 Oct 2020
How Effective is Task-Agnostic Data Augmentation for Pretrained
  Transformers?
How Effective is Task-Agnostic Data Augmentation for Pretrained Transformers?
Shayne Longpre
Yu Wang
Christopher DuBois
ViT
17
83
0
05 Oct 2020
Effective Unsupervised Domain Adaptation with Adversarially Trained
  Language Models
Effective Unsupervised Domain Adaptation with Adversarially Trained Language Models
Thuy-Trang Vu
Dinh Q. Phung
Gholamreza Haffari
8
24
0
05 Oct 2020
On Losses for Modern Language Models
On Losses for Modern Language Models
Stephane Aroca-Ouellette
Frank Rudzicz
11
33
0
04 Oct 2020
Tell Me How to Ask Again: Question Data Augmentation with Controllable
  Rewriting in Continuous Space
Tell Me How to Ask Again: Question Data Augmentation with Controllable Rewriting in Continuous Space
Dayiheng Liu
Yeyun Gong
Jie Fu
Yu Yan
Jiusheng Chen
Jiancheng Lv
Nan Duan
M. Zhou
10
37
0
04 Oct 2020
LUKE: Deep Contextualized Entity Representations with Entity-aware
  Self-attention
LUKE: Deep Contextualized Entity Representations with Entity-aware Self-attention
Ikuya Yamada
Akari Asai
Hiroyuki Shindo
Hideaki Takeda
Yuji Matsumoto
22
662
0
02 Oct 2020
Data Transfer Approaches to Improve Seq-to-Seq Retrosynthesis
Data Transfer Approaches to Improve Seq-to-Seq Retrosynthesis
Katsuhiko Ishiguro
K. Ujihara
R. Sawada
Hirotaka Akita
Masaaki Kotera
22
6
0
02 Oct 2020
Near-imperceptible Neural Linguistic Steganography via Self-Adjusting
  Arithmetic Coding
Near-imperceptible Neural Linguistic Steganography via Self-Adjusting Arithmetic Coding
Jiaming Shen
Heng Ji
Jiawei Han
13
33
0
01 Oct 2020
CoLAKE: Contextualized Language and Knowledge Embedding
CoLAKE: Contextualized Language and Knowledge Embedding
Tianxiang Sun
Yunfan Shao
Xipeng Qiu
Qipeng Guo
Yaru Hu
Xuanjing Huang
Zheng-Wei Zhang
KELM
18
181
0
01 Oct 2020
Phonemer at WNUT-2020 Task 2: Sequence Classification Using COVID
  Twitter BERT and Bagging Ensemble Technique based on Plurality Voting
Phonemer at WNUT-2020 Task 2: Sequence Classification Using COVID Twitter BERT and Bagging Ensemble Technique based on Plurality Voting
Anshul Wadhawan
14
7
0
01 Oct 2020
Examining the rhetorical capacities of neural language models
Examining the rhetorical capacities of neural language models
Zining Zhu
Chuer Pan
Mohamed Abdalla
Frank Rudzicz
28
10
0
01 Oct 2020
A Simple but Tough-to-Beat Data Augmentation Approach for Natural
  Language Understanding and Generation
A Simple but Tough-to-Beat Data Augmentation Approach for Natural Language Understanding and Generation
Dinghan Shen
Ming Zheng
Yelong Shen
Yanru Qu
Weizhu Chen
AAML
21
130
0
29 Sep 2020
The birth of Romanian BERT
The birth of Romanian BERT
Stefan Daniel Dumitrescu
Andrei-Marius Avram
S. Pyysalo
VLM
8
76
0
18 Sep 2020
Self-Supervised Meta-Learning for Few-Shot Natural Language
  Classification Tasks
Self-Supervised Meta-Learning for Few-Shot Natural Language Classification Tasks
Trapit Bansal
Rishikesh Jha
Tsendsuren Munkhdalai
Andrew McCallum
SSL
VLM
20
87
0
17 Sep 2020
GraphCodeBERT: Pre-training Code Representations with Data Flow
GraphCodeBERT: Pre-training Code Representations with Data Flow
Daya Guo
Shuo Ren
Shuai Lu
Zhangyin Feng
Duyu Tang
...
Dawn Drain
Neel Sundaresan
Jian Yin
Daxin Jiang
M. Zhou
56
1,094
0
17 Sep 2020
Code-switching pre-training for neural machine translation
Code-switching pre-training for neural machine translation
Zhen Yang
Bojie Hu
Ambyera Han
Shen Huang
Qi Ju
19
71
0
17 Sep 2020
Efficient Transformer-based Large Scale Language Representations using
  Hardware-friendly Block Structured Pruning
Efficient Transformer-based Large Scale Language Representations using Hardware-friendly Block Structured Pruning
Bingbing Li
Zhenglun Kong
Tianyun Zhang
Ji Li
Z. Li
Hang Liu
Caiwen Ding
VLM
24
64
0
17 Sep 2020
Reasoning about Goals, Steps, and Temporal Ordering with WikiHow
Reasoning about Goals, Steps, and Temporal Ordering with WikiHow
Li Zhang
Qing Lyu
Chris Callison-Burch
ReLM
LRM
16
85
0
16 Sep 2020
BERT-QE: Contextualized Query Expansion for Document Re-ranking
BERT-QE: Contextualized Query Expansion for Document Re-ranking
Zhi Zheng
Kai Hui
Ben He
Xianpei Han
Le Sun
Andrew Yates
19
93
0
15 Sep 2020
It's Not Just Size That Matters: Small Language Models Are Also Few-Shot
  Learners
It's Not Just Size That Matters: Small Language Models Are Also Few-Shot Learners
Timo Schick
Hinrich Schütze
22
953
0
15 Sep 2020
MLMLM: Link Prediction with Mean Likelihood Masked Language Model
MLMLM: Link Prediction with Mean Likelihood Masked Language Model
Louis Clouâtre
P. Trempe
Amal Zouaq
Sarath Chandar
19
43
0
15 Sep 2020
Learning an Effective Context-Response Matching Model with
  Self-Supervised Tasks for Retrieval-based Dialogues
Learning an Effective Context-Response Matching Model with Self-Supervised Tasks for Retrieval-based Dialogues
Ruijian Xu
Chongyang Tao
Daxin Jiang
Xueliang Zhao
Dongyan Zhao
Rui Yan
24
70
0
14 Sep 2020
On Robustness and Bias Analysis of BERT-based Relation Extraction
On Robustness and Bias Analysis of BERT-based Relation Extraction
Luoqiu Li
Xiang Chen
Hongbin Ye
Zhen Bi
Shumin Deng
Ningyu Zhang
Huajun Chen
26
18
0
14 Sep 2020
Probabilistic Predictions of People Perusing: Evaluating Metrics of
  Language Model Performance for Psycholinguistic Modeling
Probabilistic Predictions of People Perusing: Evaluating Metrics of Language Model Performance for Psycholinguistic Modeling
Sophie Hao
S. Mendelsohn
Rachel Sterneck
Randi Martinez
Robert Frank
6
46
0
08 Sep 2020
Adversarial Watermarking Transformer: Towards Tracing Text Provenance
  with Data Hiding
Adversarial Watermarking Transformer: Towards Tracing Text Provenance with Data Hiding
Sahar Abdelnabi
Mario Fritz
WaLM
18
143
0
07 Sep 2020
A Comparison of Pre-trained Vision-and-Language Models for Multimodal
  Representation Learning across Medical Images and Reports
A Comparison of Pre-trained Vision-and-Language Models for Multimodal Representation Learning across Medical Images and Reports
Yikuan Li
Hanyin Wang
Yuan Luo
6
63
0
03 Sep 2020
ASTRAL: Adversarial Trained LSTM-CNN for Named Entity Recognition
ASTRAL: Adversarial Trained LSTM-CNN for Named Entity Recognition
Jiuniu Wang
Wenjia Xu
Xingyu Fu
Guangluan Xu
Yirong Wu
18
57
0
02 Sep 2020
A Survey of Evaluation Metrics Used for NLG Systems
A Survey of Evaluation Metrics Used for NLG Systems
Ananya B. Sai
Akash Kumar Mohankumar
Mitesh M. Khapra
ELM
25
228
0
27 Aug 2020
Conceptualized Representation Learning for Chinese Biomedical Text
  Mining
Conceptualized Representation Learning for Chinese Biomedical Text Mining
Ningyu Zhang
Qianghuai Jia
Kangping Yin
Liang Dong
Feng Gao
Nengwei Hua
OOD
34
65
0
25 Aug 2020
Hybrid Ranking Network for Text-to-SQL
Hybrid Ranking Network for Text-to-SQL
Qin Lyu
K. Chakrabarti
Shobhit Hathi
Souvik Kundu
Jianwen Zhang
Zheng Chen
AIMat
11
83
0
11 Aug 2020
Distilling the Knowledge of BERT for Sequence-to-Sequence ASR
Distilling the Knowledge of BERT for Sequence-to-Sequence ASR
Hayato Futami
H. Inaguma
Sei Ueno
Masato Mimura
S. Sakai
Tatsuya Kawahara
19
50
0
09 Aug 2020
SemEval-2020 Task 8: Memotion Analysis -- The Visuo-Lingual Metaphor!
SemEval-2020 Task 8: Memotion Analysis -- The Visuo-Lingual Metaphor!
Chhavi Sharma
Deepesh Bhageria
W. Scott
Srinivas Pykl
A. Das
Tanmoy Chakraborty
Viswanath Pulabaigari
Björn Gambäck
20
166
0
09 Aug 2020
ConvBERT: Improving BERT with Span-based Dynamic Convolution
ConvBERT: Improving BERT with Span-based Dynamic Convolution
Zihang Jiang
Weihao Yu
Daquan Zhou
Yunpeng Chen
Jiashi Feng
Shuicheng Yan
32
156
0
06 Aug 2020
Emotion Correlation Mining Through Deep Learning Models on Natural
  Language Text
Emotion Correlation Mining Through Deep Learning Models on Natural Language Text
Xinzhi Wang
Luyao Kou
V. Sugumaran
Xiangfeng Luo
Hui Zhang
25
62
0
28 Jul 2020
FiSSA at SemEval-2020 Task 9: Fine-tuned For Feelings
FiSSA at SemEval-2020 Task 9: Fine-tuned For Feelings
Bertelt Braaksma
R. Scholtens
Stan van Suijlekom
Remy Wang
A. Ustun
15
3
0
24 Jul 2020
OnlineAugment: Online Data Augmentation with Less Domain Knowledge
OnlineAugment: Online Data Augmentation with Less Domain Knowledge
Zhiqiang Tang
Yunhe Gao
Leonid Karlinsky
P. Sattigeri
Rogerio Feris
Dimitris N. Metaxas
19
56
0
17 Jul 2020
Compositional Generalization in Semantic Parsing: Pre-training vs.
  Specialized Architectures
Compositional Generalization in Semantic Parsing: Pre-training vs. Specialized Architectures
Daniel Furrer
Marc van Zee
Nathan Scales
Nathanael Scharli
CoGe
8
113
0
17 Jul 2020
Investigating Pretrained Language Models for Graph-to-Text Generation
Investigating Pretrained Language Models for Graph-to-Text Generation
Leonardo F. R. Ribeiro
Martin Schmitt
Hinrich Schütze
Iryna Gurevych
19
215
0
16 Jul 2020
Learning Reasoning Strategies in End-to-End Differentiable Proving
Learning Reasoning Strategies in End-to-End Differentiable Proving
Pasquale Minervini
Sebastian Riedel
Pontus Stenetorp
Edward Grefenstette
Tim Rocktaschel
LRM
39
96
0
13 Jul 2020
Sparse Graph to Sequence Learning for Vision Conditioned Long Textual
  Sequence Generation
Sparse Graph to Sequence Learning for Vision Conditioned Long Textual Sequence Generation
Aditya Mogadala
Marius Mosbach
Dietrich Klakow
VLM
72
0
0
12 Jul 2020
TERA: Self-Supervised Learning of Transformer Encoder Representation for
  Speech
TERA: Self-Supervised Learning of Transformer Encoder Representation for Speech
Andy T. Liu
Shang-Wen Li
Hung-yi Lee
SSL
48
356
0
12 Jul 2020
DeepSinger: Singing Voice Synthesis with Data Mined From the Web
DeepSinger: Singing Voice Synthesis with Data Mined From the Web
Yi Ren
Xu Tan
Tao Qin
Jian Luan
Zhou Zhao
Tie-Yan Liu
28
73
0
09 Jul 2020
Continual BERT: Continual Learning for Adaptive Extractive Summarization
  of COVID-19 Literature
Continual BERT: Continual Learning for Adaptive Extractive Summarization of COVID-19 Literature
Jongjin Park
CLL
20
15
0
07 Jul 2020
DAPPLE: A Pipelined Data Parallel Approach for Training Large Models
DAPPLE: A Pipelined Data Parallel Approach for Training Large Models
Shiqing Fan
Yi Rong
Chen Meng
Zongyan Cao
Siyu Wang
...
Jun Yang
Lixue Xia
Lansong Diao
Xiaoyong Liu
Wei Lin
21
232
0
02 Jul 2020
Previous
123...171819202122
Next