ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2026 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1907.12009
  4. Cited By
Representation Degeneration Problem in Training Natural Language
  Generation Models

Representation Degeneration Problem in Training Natural Language Generation Models

International Conference on Learning Representations (ICLR), 2019
28 July 2019
Jun Gao
Di He
Xu Tan
Tao Qin
Liwei Wang
Tie-Yan Liu
ArXiv (abs)PDFHTML

Papers citing "Representation Degeneration Problem in Training Natural Language Generation Models"

50 / 162 papers shown
On Isotropy, Contextualization and Learning Dynamics of
  Contrastive-based Sentence Representation Learning
On Isotropy, Contextualization and Learning Dynamics of Contrastive-based Sentence Representation LearningAnnual Meeting of the Association for Computational Linguistics (ACL), 2022
Chenghao Xiao
Yang Long
Noura Al Moubayed
206
15
0
18 Dec 2022
HyPe: Better Pre-trained Language Model Fine-tuning with Hidden
  Representation Perturbation
HyPe: Better Pre-trained Language Model Fine-tuning with Hidden Representation PerturbationAnnual Meeting of the Association for Computational Linguistics (ACL), 2022
Hongyi Yuan
Zheng Yuan
Chuanqi Tan
Fei Huang
Songfang Huang
235
19
0
17 Dec 2022
Reliable Measures of Spread in High Dimensional Latent Spaces
Reliable Measures of Spread in High Dimensional Latent SpacesInternational Conference on Machine Learning (ICML), 2022
Anna C. Marbut
Katy McKinney-Bock
Travis J. Wheeler
289
3
0
15 Dec 2022
Self-supervised Trajectory Representation Learning with Temporal
  Regularities and Travel Semantics
Self-supervised Trajectory Representation Learning with Temporal Regularities and Travel SemanticsIEEE International Conference on Data Engineering (ICDE), 2022
Jiawei Jiang
Dayan Pan
Houxing Ren
Xiaohan Jiang
Chao Li
Jingyuan Wang
AI4TS
281
116
0
17 Nov 2022
Evade the Trap of Mediocrity: Promoting Diversity and Novelty in Text
  Generation via Concentrating Attention
Evade the Trap of Mediocrity: Promoting Diversity and Novelty in Text Generation via Concentrating AttentionConference on Empirical Methods in Natural Language Processing (EMNLP), 2022
Wenhao Li
Xiaoyuan Yi
Jinyi Hu
Maosong Sun
Xing Xie
241
2
0
14 Nov 2022
Reconciliation of Pre-trained Models and Prototypical Neural Networks in
  Few-shot Named Entity Recognition
Reconciliation of Pre-trained Models and Prototypical Neural Networks in Few-shot Named Entity RecognitionConference on Empirical Methods in Natural Language Processing (EMNLP), 2022
Youcheng Huang
Wenqiang Lei
Jie Fu
Jiancheng Lv
169
3
0
07 Nov 2022
Optimizing text representations to capture (dis)similarity between
  political parties
Optimizing text representations to capture (dis)similarity between political partiesConference on Computational Natural Language Learning (CoNLL), 2022
Tanise Ceron
Nico Blokker
Sebastian Padó
135
7
0
21 Oct 2022
Kernel-Whitening: Overcome Dataset Bias with Isotropic Sentence
  Embedding
Kernel-Whitening: Overcome Dataset Bias with Isotropic Sentence EmbeddingConference on Empirical Methods in Natural Language Processing (EMNLP), 2022
Songyang Gao
Jiajun Sun
Tao Gui
Xuanjing Huang
154
11
0
14 Oct 2022
ContraCLM: Contrastive Learning For Causal Language Model
ContraCLM: Contrastive Learning For Causal Language ModelAnnual Meeting of the Association for Computational Linguistics (ACL), 2022
Nihal Jain
Dejiao Zhang
Wasi Uddin Ahmad
Zijian Wang
Feng Nan
...
Ramesh Nallapati
Baishakhi Ray
Parminder Bhatia
Xiaofei Ma
Bing Xiang
259
22
0
03 Oct 2022
Prompt Combines Paraphrase: Teaching Pre-trained Models to Understand
  Rare Biomedical Words
Prompt Combines Paraphrase: Teaching Pre-trained Models to Understand Rare Biomedical WordsInternational Conference on Computational Linguistics (COLING), 2022
Hao Wang
Chi-Liang Liu
Nuwa Xi
Sendong Zhao
Meizhi Ju
Shiwei Zhang
Ziheng Zhang
Yefeng Zheng
Bing Qin
Ting Liu
VLMAAMLLM&MA
209
7
0
14 Sep 2022
Pre-Training a Graph Recurrent Network for Language Representation
Pre-Training a Graph Recurrent Network for Language Representation
Yile Wang
Linyi Yang
Zhiyang Teng
M. Zhou
Yue Zhang
GNN
241
1
0
08 Sep 2022
Analyzing Transformers in Embedding Space
Analyzing Transformers in Embedding SpaceAnnual Meeting of the Association for Computational Linguistics (ACL), 2022
Guy Dar
Mor Geva
Ankit Gupta
Jonathan Berant
335
124
0
06 Sep 2022
SimCLF: A Simple Contrastive Learning Framework for Function-level
  Binary Embeddings
SimCLF: A Simple Contrastive Learning Framework for Function-level Binary Embeddings
Ruijin Sun
Guo Shize
Jinhong Guo
Li Wei
Zhan Dazhi
Sun Meng
Zhisong Pan
177
0
0
06 Sep 2022
RLIP: Relational Language-Image Pre-training for Human-Object
  Interaction Detection
RLIP: Relational Language-Image Pre-training for Human-Object Interaction DetectionNeural Information Processing Systems (NeurIPS), 2022
Hangjie Yuan
Jianwen Jiang
Samuel Albanie
Tao Feng
Ziyuan Huang
Dong Ni
Mingqian Tang
VLM
374
76
0
05 Sep 2022
Isotropic Representation Can Improve Dense Retrieval
Isotropic Representation Can Improve Dense RetrievalPacific-Asia Conference on Knowledge Discovery and Data Mining (PAKDD), 2022
Euna Jung
J. Park
Jaekeol Choi
Sungyoon Kim
Wonjong Rhee
OOD
235
7
0
01 Sep 2022
Addressing Token Uniformity in Transformers via Singular Value
  Transformation
Addressing Token Uniformity in Transformers via Singular Value TransformationConference on Uncertainty in Artificial Intelligence (UAI), 2022
Hanqi Yan
Lin Gui
Wenjie Li
Yulan He
194
16
0
24 Aug 2022
Mere Contrastive Learning for Cross-Domain Sentiment Analysis
Mere Contrastive Learning for Cross-Domain Sentiment AnalysisInternational Conference on Computational Linguistics (COLING), 2022
Yun Luo
Fang Guo
Zihan Liu
Yue Zhang
155
18
0
18 Aug 2022
LLM.int8(): 8-bit Matrix Multiplication for Transformers at Scale
LLM.int8(): 8-bit Matrix Multiplication for Transformers at Scale
Tim Dettmers
M. Lewis
Younes Belkada
Luke Zettlemoyer
MQ
478
843
0
15 Aug 2022
Outliers Dimensions that Disrupt Transformers Are Driven by Frequency
Outliers Dimensions that Disrupt Transformers Are Driven by FrequencyConference on Empirical Methods in Natural Language Processing (EMNLP), 2022
Giovanni Puccetti
Anna Rogers
Aleksandr Drozd
F. Dell’Orletta
540
55
0
23 May 2022
Fine-tuning Pre-trained Language Models for Few-shot Intent Detection:
  Supervised Pre-training and Isotropization
Fine-tuning Pre-trained Language Models for Few-shot Intent Detection: Supervised Pre-training and IsotropizationNorth American Chapter of the Association for Computational Linguistics (NAACL), 2022
Haode Zhang
Haowen Liang
Yuwei Zhang
Li-Ming Zhan
Xiao-Ming Wu
Xiaolei Lu
Albert Y. S. Lam
232
34
0
15 May 2022
Label Anchored Contrastive Learning for Language Understanding
Label Anchored Contrastive Learning for Language UnderstandingNorth American Chapter of the Association for Computational Linguistics (NAACL), 2022
Zhenyu Zhang
Yuming Zhao
Meng Chen
Xiaodong He
185
17
0
26 Apr 2022
Reprint: a randomized extrapolation based on principal components for
  data augmentation
Reprint: a randomized extrapolation based on principal components for data augmentationSocial Science Research Network (SSRN), 2022
Jiale Wei
Qiyuan Chen
Pai Peng
Benjamin Guedj
Le Li
192
3
0
26 Apr 2022
A Token-level Contrastive Framework for Sign Language Translation
A Token-level Contrastive Framework for Sign Language TranslationIEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP), 2022
Biao Fu
Peigen Ye
Liang Zhang
Pei-Ju Yu
Cong Hu
Yidong Chen
X. Shi
SLR
208
17
0
11 Apr 2022
CoCoSoDa: Effective Contrastive Learning for Code Search
CoCoSoDa: Effective Contrastive Learning for Code SearchInternational Conference on Software Engineering (ICSE), 2022
Ensheng Shi
Yanlin Wang
Wenchao Gu
Lun Du
Hongyu Zhang
Shi Han
Dongmei Zhang
Hongbin Sun
322
65
0
07 Apr 2022
The Principle of Diversity: Training Stronger Vision Transformers Calls
  for Reducing All Levels of Redundancy
The Principle of Diversity: Training Stronger Vision Transformers Calls for Reducing All Levels of RedundancyComputer Vision and Pattern Recognition (CVPR), 2022
Tianlong Chen
Zhenyu Zhang
Yu Cheng
Ahmed Hassan Awadallah
Zinan Lin
ViT
260
49
0
12 Mar 2022
Mind the Gap: Understanding the Modality Gap in Multi-modal Contrastive
  Representation Learning
Mind the Gap: Understanding the Modality Gap in Multi-modal Contrastive Representation LearningNeural Information Processing Systems (NeurIPS), 2022
Weixin Liang
Yuhui Zhang
Yongchan Kwon
Serena Yeung
James Zou
VLM
455
600
0
03 Mar 2022
A Simple but Effective Pluggable Entity Lookup Table for Pre-trained
  Language Models
A Simple but Effective Pluggable Entity Lookup Table for Pre-trained Language ModelsAnnual Meeting of the Association for Computational Linguistics (ACL), 2022
Deming Ye
Yankai Lin
Peng Li
Maosong Sun
Zhiyuan Liu
KELM
214
11
0
27 Feb 2022
Exploring the Impact of Negative Samples of Contrastive Learning: A Case
  Study of Sentence Embedding
Exploring the Impact of Negative Samples of Contrastive Learning: A Case Study of Sentence EmbeddingFindings (Findings), 2022
Rui Cao
Yihao Wang
Y. Liang
Ling Gao
Jie Zheng
Jie Ren
Zheng Wang
307
41
0
26 Feb 2022
PromptBERT: Improving BERT Sentence Embeddings with Prompts
PromptBERT: Improving BERT Sentence Embeddings with PromptsConference on Empirical Methods in Natural Language Processing (EMNLP), 2022
Ting Jiang
Jian Jiao
Shaohan Huang
Zi-qiang Zhang
Deqing Wang
Fuzhen Zhuang
Furu Wei
Haizhen Huang
Liangjie Zhang
Qi Zhang
233
149
0
12 Jan 2022
Frequency-Aware Contrastive Learning for Neural Machine Translation
Frequency-Aware Contrastive Learning for Neural Machine TranslationAAAI Conference on Artificial Intelligence (AAAI), 2021
Tong Zhang
Wei Ye
Baosong Yang
Long Zhang
Xingzhang Ren
Dayiheng Liu
Jinan Sun
Shikun Zhang
Haibo Zhang
Wen Zhao
183
34
0
29 Dec 2021
A Survey of Visual Transformers
A Survey of Visual TransformersIEEE Transactions on Neural Networks and Learning Systems (TNNLS), 2021
Yang Liu
Yao Zhang
Yixin Wang
Feng Hou
Jin Yuan
Jiang Tian
Yang Zhang
Peng Wang
Jianping Fan
Zhiqiang He
3DGSViT
473
487
0
11 Nov 2021
Leveraging Advantages of Interactive and Non-Interactive Models for
  Vector-Based Cross-Lingual Information Retrieval
Leveraging Advantages of Interactive and Non-Interactive Models for Vector-Based Cross-Lingual Information Retrieval
Linlong Xu
Baosong Yang
Xiaoyu Lv
Tianchi Bi
Dayiheng Liu
Haibo Zhang
151
7
0
03 Nov 2021
Dict-BERT: Enhancing Language Model Pre-training with Dictionary
Dict-BERT: Enhancing Language Model Pre-training with Dictionary
Wenhao Yu
Chenguang Zhu
Yuwei Fang
Donghan Yu
Shuohang Wang
Yichong Xu
Michael Zeng
Meng Jiang
383
68
0
13 Oct 2021
An Isotropy Analysis in the Multilingual BERT Embedding Space
An Isotropy Analysis in the Multilingual BERT Embedding SpaceFindings (Findings), 2021
S. Rajaee
Mohammad Taher Pilehvar
246
36
0
09 Oct 2021
Text analysis and deep learning: A network approach
Text analysis and deep learning: A network approach
Ingo Marquart
171
0
0
08 Oct 2021
On Isotropy Calibration of Transformers
On Isotropy Calibration of TransformersFirst Workshop on Insights from Negative Results in NLP (Insights), 2021
Yue Ding
Karolis Martinkus
Damian Pascual
Simon Clematide
Roger Wattenhofer
131
1
0
27 Sep 2021
How Does Fine-tuning Affect the Geometry of Embedding Space: A Case
  Study on Isotropy
How Does Fine-tuning Affect the Geometry of Embedding Space: A Case Study on IsotropyConference on Empirical Methods in Natural Language Processing (EMNLP), 2021
S. Rajaee
Mohammad Taher Pilehvar
261
27
0
10 Sep 2021
All Bark and No Bite: Rogue Dimensions in Transformer Language Models
  Obscure Representational Quality
All Bark and No Bite: Rogue Dimensions in Transformer Language Models Obscure Representational QualityConference on Empirical Methods in Natural Language Processing (EMNLP), 2021
William Timkey
Marten van Schijndel
504
133
0
09 Sep 2021
Rare Tokens Degenerate All Tokens: Improving Neural Text Generation via
  Adaptive Gradient Gating for Rare Token Embeddings
Rare Tokens Degenerate All Tokens: Improving Neural Text Generation via Adaptive Gradient Gating for Rare Token EmbeddingsAnnual Meeting of the Association for Computational Linguistics (ACL), 2021
Sangwon Yu
Jongyoon Song
Heeseung Kim
SeongEun Lee
Woo-Jong Ryu
Sung-Hoon Yoon
395
39
0
07 Sep 2021
IsoScore: Measuring the Uniformity of Embedding Space Utilization
IsoScore: Measuring the Uniformity of Embedding Space Utilization
William Rudman
Nate Gillman
T. Rayne
Carsten Eickhoff
223
36
0
16 Aug 2021
Language Models as Zero-shot Visual Semantic Learners
Language Models as Zero-shot Visual Semantic Learners
Yue Jiao
Jonathon S. Hare
Adam Prugel-Bennett
VLM
109
1
0
26 Jul 2021
Noisy Training Improves E2E ASR for the Edge
Noisy Training Improves E2E ASR for the Edge
Dilin Wang
Yuan Shangguan
Haichuan Yang
P. Chuang
Jiatong Zhou
Meng Li
Ganesh Venkatesh
Ozlem Kalinli
Vikas Chandra
231
4
0
09 Jul 2021
A Cluster-based Approach for Improving Isotropy in Contextual Embedding
  Space
A Cluster-based Approach for Improving Isotropy in Contextual Embedding SpaceAnnual Meeting of the Association for Computational Linguistics (ACL), 2021
S. Rajaee
Mohammad Taher Pilehvar
162
44
0
02 Jun 2021
Rejuvenating Low-Frequency Words: Making the Most of Parallel Data in
  Non-Autoregressive Translation
Rejuvenating Low-Frequency Words: Making the Most of Parallel Data in Non-Autoregressive TranslationAnnual Meeting of the Association for Computational Linguistics (ACL), 2021
Liang Ding
Longyue Wang
Xuebo Liu
Yang Li
Dacheng Tao
Zhaopeng Tu
214
50
0
02 Jun 2021
ConSERT: A Contrastive Framework for Self-Supervised Sentence
  Representation Transfer
ConSERT: A Contrastive Framework for Self-Supervised Sentence Representation TransferAnnual Meeting of the Association for Computational Linguistics (ACL), 2021
Yuanmeng Yan
Rumei Li
Sirui Wang
Fuzheng Zhang
Wei Wu
Weiran Xu
SSL
282
617
0
25 May 2021
Vision Transformers with Patch Diversification
Vision Transformers with Patch Diversification
Chengyue Gong
Dilin Wang
Meng Li
Vikas Chandra
Qiang Liu
ViT
257
68
0
26 Apr 2021
Low Anisotropy Sense Retrofitting (LASeR) : Towards Isotropic and Sense
  Enriched Representations
Low Anisotropy Sense Retrofitting (LASeR) : Towards Isotropic and Sense Enriched RepresentationsWorkshop on Knowledge Extraction and Integration for Deep Learning Architectures; Deep Learning Inside Out (DEELIO), 2021
Geetanjali Bihani
Julia Taylor Rayz
172
13
0
22 Apr 2021
SimCSE: Simple Contrastive Learning of Sentence Embeddings
SimCSE: Simple Contrastive Learning of Sentence EmbeddingsConference on Empirical Methods in Natural Language Processing (EMNLP), 2021
Tianyu Gao
Xingcheng Yao
Danqi Chen
AILawSSL
829
4,055
0
18 Apr 2021
Learning to Remove: Towards Isotropic Pre-trained BERT Embedding
Learning to Remove: Towards Isotropic Pre-trained BERT EmbeddingInternational Conference on Artificial Neural Networks (ICANN), 2021
Y. Liang
Rui Cao
Jie Zheng
Jie Ren
Ling Gao
SSL
433
32
0
12 Apr 2021
Whitening Sentence Representations for Better Semantics and Faster
  Retrieval
Whitening Sentence Representations for Better Semantics and Faster Retrieval
Jianlin Su
Jiarun Cao
Weijie Liu
Yangyiwen Ou
300
338
0
29 Mar 2021
Previous
1234
Next
Page 3 of 4