ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2010.02648
  4. Cited By
On the Sub-Layer Functionalities of Transformer Decoder

On the Sub-Layer Functionalities of Transformer Decoder

Findings (Findings), 2020
6 October 2020
Yilin Yang
Longyue Wang
Shuming Shi
Prasad Tadepalli
Stefan Lee
Zhaopeng Tu
ArXiv (abs)PDFHTML

Papers citing "On the Sub-Layer Functionalities of Transformer Decoder"

13 / 13 papers shown
Title
Decoder-only Streaming Transformer for Simultaneous Translation
Decoder-only Streaming Transformer for Simultaneous Translation
Shoutao Guo
Shaolei Zhang
Yang Feng
220
13
0
06 Jun 2024
Salute the Classic: Revisiting Challenges of Machine Translation in the
  Age of Large Language Models
Salute the Classic: Revisiting Challenges of Machine Translation in the Age of Large Language ModelsTransactions of the Association for Computational Linguistics (TACL), 2024
Jianhui Pang
Fanghua Ye
Longyue Wang
Dian Yu
Derek F. Wong
Shuming Shi
Zhaopeng Tu
ALM
172
20
0
16 Jan 2024
Understanding LLMs: A Comprehensive Overview from Training to Inference
Understanding LLMs: A Comprehensive Overview from Training to Inference
Yi-Hsueh Liu
Haoyang He
Tianle Han
Xu-Yao Zhang
Mengyuan Liu
...
Xiaoyan Cai
Tuo Zhang
Ning Qiang
Tianming Liu
Bao Ge
SyDa
370
120
0
04 Jan 2024
Predicting Human Translation Difficulty with Neural Machine Translation
Predicting Human Translation Difficulty with Neural Machine Translation
Zheng Wei Lim
Ekaterina Vylomova
Charles Kemp
Trevor Cohn
171
1
0
19 Dec 2023
ChatGPT-Like Large-Scale Foundation Models for Prognostics and Health
  Management: A Survey and Roadmaps
ChatGPT-Like Large-Scale Foundation Models for Prognostics and Health Management: A Survey and RoadmapsReliability Engineering & System Safety (Reliab. Eng. Syst. Saf.), 2023
Yanfang Li
Huan Wang
Muxia Sun
LM&MAAI4TSAI4CE
287
83
0
10 May 2023
Greedy Ordering of Layer Weight Matrices in Transformers Improves Translation
Elicia Ye
150
1
0
04 Feb 2023
Unveiling the Black Box of PLMs with Semantic Anchors: Towards
  Interpretable Neural Semantic Parsing
Unveiling the Black Box of PLMs with Semantic Anchors: Towards Interpretable Neural Semantic ParsingAAAI Conference on Artificial Intelligence (AAAI), 2022
L. Nie
Jiu Sun
Yanlin Wang
Lun Du
Lei Hou
Juanzi Li
Shi Han
Dongmei Zhang
Jidong Zhai
164
7
0
04 Oct 2022
Can Transformer be Too Compositional? Analysing Idiom Processing in
  Neural Machine Translation
Can Transformer be Too Compositional? Analysing Idiom Processing in Neural Machine TranslationAnnual Meeting of the Association for Computational Linguistics (ACL), 2022
Verna Dankers
Christopher G. Lucas
Ivan Titov
163
42
0
30 May 2022
Transformer Feed-Forward Layers Build Predictions by Promoting Concepts
  in the Vocabulary Space
Transformer Feed-Forward Layers Build Predictions by Promoting Concepts in the Vocabulary SpaceConference on Empirical Methods in Natural Language Processing (EMNLP), 2022
Mor Geva
Avi Caciularu
Ke Wang
Yoav Goldberg
KELM
505
449
0
28 Mar 2022
Improving Multilingual Translation by Representation and Gradient
  Regularization
Improving Multilingual Translation by Representation and Gradient RegularizationConference on Empirical Methods in Natural Language Processing (EMNLP), 2021
Yilin Yang
Akiko Eriguchi
Alexandre Muzio
Prasad Tadepalli
Stefan Lee
Hany Hassan
117
41
0
10 Sep 2021
A Survey of Transformers
A Survey of TransformersAI Open (AO), 2021
Tianyang Lin
Yuxin Wang
Xiangyang Liu
Xipeng Qiu
ViT
349
1,344
0
08 Jun 2021
Context-Aware Cross-Attention for Non-Autoregressive Translation
Context-Aware Cross-Attention for Non-Autoregressive TranslationInternational Conference on Computational Linguistics (COLING), 2020
Liang Ding
Longyue Wang
Di Wu
Dacheng Tao
Zhaopeng Tu
144
41
0
02 Nov 2020
On the Sparsity of Neural Machine Translation Models
On the Sparsity of Neural Machine Translation ModelsConference on Empirical Methods in Natural Language Processing (EMNLP), 2020
Yong Wang
Longyue Wang
Victor O.K. Li
Zhaopeng Tu
MoE
116
11
0
06 Oct 2020
1