ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2212.01853
  4. Cited By
Toward Efficient Language Model Pretraining and Downstream Adaptation
  via Self-Evolution: A Case Study on SuperGLUE

Toward Efficient Language Model Pretraining and Downstream Adaptation via Self-Evolution: A Case Study on SuperGLUE

4 December 2022
Qihuang Zhong
Liang Ding
Yibing Zhan
Yu Qiao
Yonggang Wen
Li Shen
Juhua Liu
Baosheng Yu
Bo Du
Yixin Chen
Xinbo Gao
Steven C. H. Hoi
Xiaoou Tang
Dacheng Tao
    VLMELM
ArXiv (abs)PDFHTML

Papers citing "Toward Efficient Language Model Pretraining and Downstream Adaptation via Self-Evolution: A Case Study on SuperGLUE"

21 / 21 papers shown
Title
Survey of NLU Benchmarks Diagnosing Linguistic Phenomena: Why not Standardize Diagnostics Benchmarks?
Survey of NLU Benchmarks Diagnosing Linguistic Phenomena: Why not Standardize Diagnostics Benchmarks?
Khloud Al Jallad
Nada Ghneim
Ghaida Rebdawi
LM&MAELM
164
0
0
27 Jul 2025
What Are They Filtering Out? An Experimental Benchmark of Filtering Strategies for Harm Reduction in Pretraining Datasets
What Are They Filtering Out? An Experimental Benchmark of Filtering Strategies for Harm Reduction in Pretraining Datasets
Marco Antonio Stranisci
Christian Hardmeier
354
2
0
17 Feb 2025
Self-Evolution Knowledge Distillation for LLM-based Machine Translation
Self-Evolution Knowledge Distillation for LLM-based Machine TranslationInternational Conference on Computational Linguistics (COLING), 2024
Yuncheng Song
Liang Ding
Changtong Zan
Shujian Huang
325
0
0
19 Dec 2024
CLOCR-C: Context Leveraging OCR Correction with Pre-trained Language Models
CLOCR-C: Context Leveraging OCR Correction with Pre-trained Language Models
Jonathan Bourne
317
6
0
30 Aug 2024
Open Sentence Embeddings for Portuguese with the Serafim PT* encoders
  family
Open Sentence Embeddings for Portuguese with the Serafim PT* encoders familyPortuguese Conference on Artificial Intelligence (EPIA), 2024
Luís Gomes
António Branco
Joao Silva
João Rodrigues
Rodrigo Santos
3DV
122
5
0
28 Jul 2024
Evaluating Implicit Bias in Large Language Models by Attacking From a Psychometric Perspective
Evaluating Implicit Bias in Large Language Models by Attacking From a Psychometric Perspective
Yuchen Wen
Keping Bi
Wei Chen
Jiafeng Guo
Xueqi Cheng
409
6
0
20 Jun 2024
Achieving >97% on GSM8K: Deeply Understanding the Problems Makes LLMs Better Solvers for Math Word Problems
Achieving >97% on GSM8K: Deeply Understanding the Problems Makes LLMs Better Solvers for Math Word Problems
Qihuang Zhong
Kang Wang
Ziyang Xu
Juhua Liu
Liang Ding
Bo Du
LRMAIMat
401
6
0
23 Apr 2024
Take Care of Your Prompt Bias! Investigating and Mitigating Prompt Bias
  in Factual Knowledge Extraction
Take Care of Your Prompt Bias! Investigating and Mitigating Prompt Bias in Factual Knowledge ExtractionInternational Conference on Language Resources and Evaluation (LREC), 2024
Ziyang Xu
Keqin Peng
Liang Ding
Dacheng Tao
Xiliang Lu
138
18
0
15 Mar 2024
Fostering the Ecosystem of Open Neural Encoders for Portuguese with
  Albertina PT* Family
Fostering the Ecosystem of Open Neural Encoders for Portuguese with Albertina PT* Family
Rodrigo Santos
João Rodrigues
Luís Gomes
Joao Silva
António Branco
Henrique Lopes Cardoso
T. Osório
Bernardo Leite
154
12
0
04 Mar 2024
Zero-Shot Sharpness-Aware Quantization for Pre-trained Language Models
Zero-Shot Sharpness-Aware Quantization for Pre-trained Language Models
Miaoxi Zhu
Qihuang Zhong
Li Shen
Liang Ding
Juhua Liu
Bo Du
Dacheng Tao
MQVLM
136
2
0
20 Oct 2023
Diversifying the Mixture-of-Experts Representation for Language Models
  with Orthogonal Optimizer
Diversifying the Mixture-of-Experts Representation for Language Models with Orthogonal Optimizer
Boan Liu
Liang Ding
Li Shen
Keqin Peng
Yu Cao
Dazhao Cheng
Dacheng Tao
MoE
154
17
0
15 Oct 2023
Explaining Relation Classification Models with Semantic Extents
Explaining Relation Classification Models with Semantic Extents
Lars Klöser
André Büsgen
Philipp Kohl
Bodo Kraft
Albert Zündorf
94
0
0
04 Aug 2023
Divide, Conquer, and Combine: Mixture of Semantic-Independent Experts
  for Zero-Shot Dialogue State Tracking
Divide, Conquer, and Combine: Mixture of Semantic-Independent Experts for Zero-Shot Dialogue State TrackingAnnual Meeting of the Association for Computational Linguistics (ACL), 2023
Qingyue Wang
Liang Ding
Yanan Cao
Yibing Zhan
Zheng Lin
Zhiliang Tian
Dacheng Tao
Li Guo
MoMeMoE
166
16
0
01 Jun 2023
Self-Evolution Learning for Discriminative Language Model Pretraining
Self-Evolution Learning for Discriminative Language Model PretrainingAnnual Meeting of the Association for Computational Linguistics (ACL), 2023
Qihuang Zhong
Liang Ding
Juhua Liu
Bo Du
Dacheng Tao
171
12
0
24 May 2023
Revisiting Token Dropping Strategy in Efficient BERT Pretraining
Revisiting Token Dropping Strategy in Efficient BERT PretrainingAnnual Meeting of the Association for Computational Linguistics (ACL), 2023
Qihuang Zhong
Liang Ding
Juhua Liu
Xuebo Liu
Min Zhang
Bo Du
Dacheng Tao
VLM
126
11
0
24 May 2023
Prompt-Learning for Cross-Lingual Relation Extraction
Prompt-Learning for Cross-Lingual Relation ExtractionIEEE International Joint Conference on Neural Network (IJCNN), 2023
Chiaming Hsu
Changtong Zan
Liang Ding
Longyue Wang
Xiaoting Wang
Weifeng Liu
Fu Lin
Wenbin Hu
LRM
144
12
0
20 Apr 2023
Towards Making the Most of ChatGPT for Machine Translation
Towards Making the Most of ChatGPT for Machine TranslationConference on Empirical Methods in Natural Language Processing (EMNLP), 2023
Keqin Peng
Liang Ding
Qihuang Zhong
Li Shen
Xuebo Liu
Min Zhang
Y. Ouyang
Dacheng Tao
LRM
291
290
0
24 Mar 2023
OmniForce: On Human-Centered, Large Model Empowered and Cloud-Edge
  Collaborative AutoML System
OmniForce: On Human-Centered, Large Model Empowered and Cloud-Edge Collaborative AutoML System
Chao Xue
Wen Liu
Shunxing Xie
Zhenfang Wang
Jiaxing Li
...
Shi-Yong Chen
Yibing Zhan
Jing Zhang
Chaoyue Wang
Dacheng Tao
203
2
0
01 Mar 2023
Can ChatGPT Understand Too? A Comparative Study on ChatGPT and
  Fine-tuned BERT
Can ChatGPT Understand Too? A Comparative Study on ChatGPT and Fine-tuned BERT
Qihuang Zhong
Liang Ding
Juhua Liu
Bo Du
Dacheng Tao
AI4MH
252
277
0
19 Feb 2023
Bag of Tricks for Effective Language Model Pretraining and Downstream
  Adaptation: A Case Study on GLUE
Bag of Tricks for Effective Language Model Pretraining and Downstream Adaptation: A Case Study on GLUE
Qihuang Zhong
Liang Ding
Keqin Peng
Juhua Liu
Bo Du
Li Shen
Yibing Zhan
Dacheng Tao
VLM
102
15
0
18 Feb 2023
PANDA: Prompt Transfer Meets Knowledge Distillation for Efficient Model
  Adaptation
PANDA: Prompt Transfer Meets Knowledge Distillation for Efficient Model AdaptationIEEE Transactions on Knowledge and Data Engineering (TKDE), 2022
Qihuang Zhong
Liang Ding
Juhua Liu
Bo Du
Dacheng Tao
VLMCLL
157
50
0
22 Aug 2022
1