ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2026 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2005.00697
  4. Cited By
DeFormer: Decomposing Pre-trained Transformers for Faster Question
  Answering

DeFormer: Decomposing Pre-trained Transformers for Faster Question Answering

Annual Meeting of the Association for Computational Linguistics (ACL), 2020
2 May 2020
Qingqing Cao
H. Trivedi
A. Balasubramanian
Niranjan Balasubramanian
ArXiv (abs)PDFHTMLGithub (120★)

Papers citing "DeFormer: Decomposing Pre-trained Transformers for Faster Question Answering"

27 / 27 papers shown
Enhancing Speech Emotion Recognition with Multi-Task Learning and Dynamic Feature Fusion
Enhancing Speech Emotion Recognition with Multi-Task Learning and Dynamic Feature Fusion
Honghong Wang
Jing Deng
Fanqin Meng
Rong Zheng
99
1
0
25 Aug 2025
Comparing Neighbors Together Makes it Easy: Jointly Comparing Multiple
  Candidates for Efficient and Effective Retrieval
Comparing Neighbors Together Makes it Easy: Jointly Comparing Multiple Candidates for Efficient and Effective Retrieval
Jonghyun Song
Cheyon Jin
Wenlong Zhao
Jay Yoon Lee
359
3
0
21 May 2024
Vesper: A Compact and Effective Pretrained Model for Speech Emotion
  Recognition
Vesper: A Compact and Effective Pretrained Model for Speech Emotion RecognitionIEEE Transactions on Affective Computing (IEEE Trans. Affective Comput.), 2023
Weidong Chen
Xiaofen Xing
Peihao Chen
Xiangmin Xu
VLM
345
75
0
20 Jul 2023
A Monte Carlo Language Model Pipeline for Zero-Shot Sociopolitical Event
  Extraction
A Monte Carlo Language Model Pipeline for Zero-Shot Sociopolitical Event Extraction
Erica Cai
Brendan O'Connor
219
6
0
24 May 2023
Investigating the Role of Feed-Forward Networks in Transformers Using
  Parallel Attention and Feed-Forward Net Design
Investigating the Role of Feed-Forward Networks in Transformers Using Parallel Attention and Feed-Forward Net Design
Shashank Sonkar
Richard G. Baraniuk
229
6
0
22 May 2023
AttMEMO : Accelerating Transformers with Memoization on Big Memory
  Systems
AttMEMO : Accelerating Transformers with Memoization on Big Memory Systems
Yuan Feng
Hyeran Jeon
F. Blagojevic
Cyril Guyot
Qing Li
Dong Li
GNN
357
8
0
23 Jan 2023
Once is Enough: A Light-Weight Cross-Attention for Fast Sentence Pair
  Modeling
Once is Enough: A Light-Weight Cross-Attention for Fast Sentence Pair ModelingConference on Empirical Methods in Natural Language Processing (EMNLP), 2022
Yuanhang Yang
Shiyi Qi
Chuanyi Liu
Qifan Wang
Cuiyun Gao
Zenglin Xu
187
3
0
11 Oct 2022
RAAT: Relation-Augmented Attention Transformer for Relation Modeling in
  Document-Level Event Extraction
RAAT: Relation-Augmented Attention Transformer for Relation Modeling in Document-Level Event ExtractionNorth American Chapter of the Association for Computational Linguistics (NAACL), 2022
Yuan Liang
Zhuoxuan Jiang
Di Yin
Bo Ren
253
36
0
07 Jun 2022
Differentially Private Model Compression
Differentially Private Model CompressionNeural Information Processing Systems (NeurIPS), 2022
Fatemehsadat Mireshghallah
A. Backurs
Huseyin A. Inan
Lukas Wutschitz
Janardhan Kulkarni
SyDa
236
16
0
03 Jun 2022
Exploring Extreme Parameter Compression for Pre-trained Language Models
Exploring Extreme Parameter Compression for Pre-trained Language ModelsInternational Conference on Learning Representations (ICLR), 2022
Yuxin Ren
Benyou Wang
Lifeng Shang
Xin Jiang
Qun Liu
268
23
0
20 May 2022
Transkimmer: Transformer Learns to Layer-wise Skim
Transkimmer: Transformer Learns to Layer-wise SkimAnnual Meeting of the Association for Computational Linguistics (ACL), 2022
Yue Guan
Zhengyi Li
Jingwen Leng
Zhouhan Lin
Minyi Guo
188
43
0
15 May 2022
Enable Deep Learning on Mobile Devices: Methods, Systems, and
  Applications
Enable Deep Learning on Mobile Devices: Methods, Systems, and Applications
Han Cai
Ji Lin
Chengyue Wu
Zhijian Liu
Haotian Tang
Hanrui Wang
Ligeng Zhu
Song Han
288
138
0
25 Apr 2022
Question Generation for Evaluating Cross-Dataset Shifts in Multi-modal
  Grounding
Question Generation for Evaluating Cross-Dataset Shifts in Multi-modal Grounding
Arjun Reddy Akula
OOD
228
3
0
24 Jan 2022
Block-Skim: Efficient Question Answering for Transformer
Block-Skim: Efficient Question Answering for Transformer
Yue Guan
Zhengyi Li
Jingwen Leng
Zhouhan Lin
Minyi Guo
Yuhao Zhu
271
33
0
16 Dec 2021
VIRT: Improving Representation-based Models for Text Matching through
  Virtual Interaction
VIRT: Improving Representation-based Models for Text Matching through Virtual Interaction
Dan Li
Yang Yang
Hongyin Tang
Jingang Wang
Tong Xu
Wei Wu
Enhong Chen
266
11
0
08 Dec 2021
A Survey on Deep Learning Event Extraction: Approaches and Applications
A Survey on Deep Learning Event Extraction: Approaches and Applications
Qian Li
Jianxin Li
Shuaiyi Nie
Shiyao Cui
Hongzhi Zhang
...
Hao Peng
Shu Guo
Lihong Wang
Amin Beheshti
Philip S. Yu
359
69
0
05 Jul 2021
TR-BERT: Dynamic Token Reduction for Accelerating BERT Inference
TR-BERT: Dynamic Token Reduction for Accelerating BERT InferenceNorth American Chapter of the Association for Computational Linguistics (NAACL), 2021
Deming Ye
Yankai Lin
Yufei Huang
Maosong Sun
MQ
288
76
0
25 May 2021
Retrieval-Free Knowledge-Grounded Dialogue Response Generation with
  Adapters
Retrieval-Free Knowledge-Grounded Dialogue Response Generation with AdaptersWorkshop on Document-grounded Dialogue and Conversational Question Answering (DialDoc), 2021
Yan Xu
Etsuko Ishii
Samuel Cahyawijaya
Zihan Liu
Genta Indra Winata
Andrea Madotto
Jane Polak Scowcroft
Pascale Fung
RALM
259
47
0
13 May 2021
Adapting by Pruning: A Case Study on BERT
Adapting by Pruning: A Case Study on BERT
Yang Gao
Nicolo Colombo
Wen Wang
177
22
0
07 May 2021
Probing Classifiers: Promises, Shortcomings, and Advances
Probing Classifiers: Promises, Shortcomings, and AdvancesInternational Conference on Computational Logic (ICCL), 2021
Yonatan Belinkov
928
697
0
24 Feb 2021
Optimizing Inference Performance of Transformers on CPUs
Optimizing Inference Performance of Transformers on CPUs
D. Dice
Alex Kogan
211
21
0
12 Feb 2021
Modeling Context in Answer Sentence Selection Systems on a Latency
  Budget
Modeling Context in Answer Sentence Selection Systems on a Latency BudgetConference of the European Chapter of the Association for Computational Linguistics (EACL), 2021
Rujun Han
Luca Soldaini
Alessandro Moschitti
273
14
0
28 Jan 2021
Learning Dense Representations of Phrases at Scale
Learning Dense Representations of Phrases at ScaleAnnual Meeting of the Association for Computational Linguistics (ACL), 2020
Jinhyuk Lee
Mujeen Sung
Jaewoo Kang
Danqi Chen
RALMDMLNAI
510
128
0
23 Dec 2020
ReadOnce Transformers: Reusable Representations of Text for Transformers
ReadOnce Transformers: Reusable Representations of Text for TransformersAnnual Meeting of the Association for Computational Linguistics (ACL), 2020
Shih-Ting Lin
Ashish Sabharwal
Tushar Khot
334
3
0
24 Oct 2020
Which *BERT? A Survey Organizing Contextualized Encoders
Which *BERT? A Survey Organizing Contextualized EncodersConference on Empirical Methods in Natural Language Processing (EMNLP), 2020
Patrick Xia
Shijie Wu
Benjamin Van Durme
396
53
0
02 Oct 2020
Adding Recurrence to Pretrained Transformers for Improved Efficiency and
  Context Size
Adding Recurrence to Pretrained Transformers for Improved Efficiency and Context Size
Davis Yoshida
Allyson Ettinger
Kevin Gimpel
AI4CE
256
7
0
16 Aug 2020
Compressing Large-Scale Transformer-Based Models: A Case Study on BERT
Compressing Large-Scale Transformer-Based Models: A Case Study on BERTTransactions of the Association for Computational Linguistics (TACL), 2020
Prakhar Ganesh
Yao Chen
Xin Lou
Mohammad Ali Khan
Yifan Yang
Hassan Sajjad
Preslav Nakov
Deming Chen
Marianne Winslett
AI4CE
625
213
0
27 Feb 2020
1
Page 1 of 1