ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1704.03471
  4. Cited By
What do Neural Machine Translation Models Learn about Morphology?
v1v2v3 (latest)

What do Neural Machine Translation Models Learn about Morphology?

11 April 2017
Yonatan Belinkov
Nadir Durrani
Fahim Dalvi
Hassan Sajjad
James R. Glass
ArXiv (abs)PDFHTML

Papers citing "What do Neural Machine Translation Models Learn about Morphology?"

50 / 251 papers shown
Title
The Interpreter Understands Your Meaning: End-to-end Spoken Language
  Understanding Aided by Speech Translation
The Interpreter Understands Your Meaning: End-to-end Spoken Language Understanding Aided by Speech TranslationConference on Empirical Methods in Natural Language Processing (EMNLP), 2023
Mutian He
Philip N. Garner
307
5
0
16 May 2023
How Good are Commercial Large Language Models on African Languages?
How Good are Commercial Large Language Models on African Languages?
Jessica Ojo
Kelechi Ogueji
157
6
0
11 May 2023
Decoder-Only or Encoder-Decoder? Interpreting Language Model as a
  Regularized Encoder-Decoder
Decoder-Only or Encoder-Decoder? Interpreting Language Model as a Regularized Encoder-Decoder
Z. Fu
W. Lam
Qian Yu
Anthony Man-Cho So
Shengding Hu
Zhiyuan Liu
Nigel Collier
AuLLM
139
58
0
08 Apr 2023
Do Transformers Parse while Predicting the Masked Word?
Do Transformers Parse while Predicting the Masked Word?Conference on Empirical Methods in Natural Language Processing (EMNLP), 2023
Haoyu Zhao
A. Panigrahi
Rong Ge
Sanjeev Arora
302
39
0
14 Mar 2023
NxPlain: Web-based Tool for Discovery of Latent Concepts
NxPlain: Web-based Tool for Discovery of Latent ConceptsConference of the European Chapter of the Association for Computational Linguistics (EACL), 2023
Fahim Dalvi
Nadir Durrani
Hassan Sajjad
Tamim Jaban
Musab Husaini
Ummar Abbas
177
1
0
06 Mar 2023
On the Role of Morphological Information for Contextual Lemmatization
On the Role of Morphological Information for Contextual LemmatizationInternational Conference on Computational Logic (ICCL), 2023
Olia Toporkov
Rodrigo Agerri
285
13
0
01 Feb 2023
Evaluating Neuron Interpretation Methods of NLP Models
Evaluating Neuron Interpretation Methods of NLP ModelsNeural Information Processing Systems (NeurIPS), 2023
Yimin Fan
Fahim Dalvi
Nadir Durrani
Hassan Sajjad
241
9
0
30 Jan 2023
The Role of Interactive Visualization in Explaining (Large) NLP Models:
  from Data to Inference
The Role of Interactive Visualization in Explaining (Large) NLP Models: from Data to Inference
R. Brath
Daniel A. Keim
Johannes Knittel
Shimei Pan
Pia Sommerauer
Hendrik Strobelt
116
14
0
11 Jan 2023
Can Large Language Models Change User Preference Adversarially?
Can Large Language Models Change User Preference Adversarially?
Varshini Subhash
AAML
158
9
0
05 Jan 2023
Intent Recognition in Conversational Recommender Systems
Intent Recognition in Conversational Recommender Systems
Sahar Moradizeyveh
252
7
0
06 Dec 2022
Localization vs. Semantics: Visual Representations in Unimodal and
  Multimodal Models
Localization vs. Semantics: Visual Representations in Unimodal and Multimodal ModelsConference of the European Chapter of the Association for Computational Linguistics (EACL), 2022
Zhuowan Li
Cihang Xie
Benjamin Van Durme
Yaoyao Liu
VLMSSL
119
2
0
01 Dec 2022
Transferability Estimation Based On Principal Gradient Expectation
Transferability Estimation Based On Principal Gradient Expectation
Huiyan Qi
Lechao Cheng
Yue Yu
Yue Yu
Haijun Shan
Zunlei Feng
Yueping Jiang
202
4
0
29 Nov 2022
ConceptX: A Framework for Latent Concept Analysis
ConceptX: A Framework for Latent Concept AnalysisAAAI Conference on Artificial Intelligence (AAAI), 2022
Firoj Alam
Fahim Dalvi
Nadir Durrani
Hassan Sajjad
A. Khan
Jia Xu
113
6
0
12 Nov 2022
The Architectural Bottleneck Principle
The Architectural Bottleneck PrincipleConference on Empirical Methods in Natural Language Processing (EMNLP), 2022
Tiago Pimentel
Josef Valvoda
Niklas Stoehr
Robert Bamler
143
5
0
11 Nov 2022
Impact of Adversarial Training on Robustness and Generalizability of
  Language Models
Impact of Adversarial Training on Robustness and Generalizability of Language ModelsAnnual Meeting of the Association for Computational Linguistics (ACL), 2022
Enes Altinisik
Hassan Sajjad
Husrev Taha Sencar
Safa Messaoud
Sanjay Chawla
AAML
164
15
0
10 Nov 2022
BLOOM: A 176B-Parameter Open-Access Multilingual Language Model
BLOOM: A 176B-Parameter Open-Access Multilingual Language Model
BigScience Workshop
:
Teven Le Scao
Angela Fan
Christopher Akiki
...
Zhongli Xie
Zifan Ye
M. Bras
Younes Belkada
Thomas Wolf
VLM
808
2,715
0
09 Nov 2022
Do Charge Prediction Models Learn Legal Theory?
Do Charge Prediction Models Learn Legal Theory?Conference on Empirical Methods in Natural Language Processing (EMNLP), 2022
Zhenwei An
Quzhe Huang
Cong Jiang
Yansong Feng
Dongyan Zhao
ELMAILaw
132
8
0
31 Oct 2022
The Better Your Syntax, the Better Your Semantics? Probing Pretrained
  Language Models for the English Comparative Correlative
The Better Your Syntax, the Better Your Semantics? Probing Pretrained Language Models for the English Comparative CorrelativeConference on Empirical Methods in Natural Language Processing (EMNLP), 2022
Leonie Weissweiler
Valentin Hofmann
Abdullatif Köksal
Hinrich Schütze
129
42
0
24 Oct 2022
On the Transformation of Latent Space in Fine-Tuned NLP Models
On the Transformation of Latent Space in Fine-Tuned NLP ModelsConference on Empirical Methods in Natural Language Processing (EMNLP), 2022
Nadir Durrani
Hassan Sajjad
Fahim Dalvi
Firoj Alam
241
20
0
23 Oct 2022
Understanding Domain Learning in Language Models Through Subpopulation
  Analysis
Understanding Domain Learning in Language Models Through Subpopulation AnalysisBlackboxNLP Workshop on Analyzing and Interpreting Neural Networks for NLP (BlackboxNLP), 2022
Zheng Zhao
Yftah Ziser
Shay B. Cohen
169
7
0
22 Oct 2022
Probing with Noise: Unpicking the Warp and Weft of Embeddings
Probing with Noise: Unpicking the Warp and Weft of EmbeddingsBlackboxNLP Workshop on Analyzing and Interpreting Neural Networks for NLP (BlackboxNLP), 2022
Filip Klubicka
John D. Kelleher
169
4
0
21 Oct 2022
Post-hoc analysis of Arabic transformer models
Post-hoc analysis of Arabic transformer modelsBlackboxNLP Workshop on Analyzing and Interpreting Neural Networks for NLP (BlackboxNLP), 2022
Ahmed Abdelali
Nadir Durrani
Fahim Dalvi
Hassan Sajjad
108
1
0
18 Oct 2022
Measures of Information Reflect Memorization Patterns
Measures of Information Reflect Memorization PatternsNeural Information Processing Systems (NeurIPS), 2022
Rachit Bansal
Danish Pruthi
Yonatan Belinkov
336
12
0
17 Oct 2022
Predicting Fine-Tuning Performance with Probing
Predicting Fine-Tuning Performance with ProbingConference on Empirical Methods in Natural Language Processing (EMNLP), 2022
Zining Zhu
Soroosh Shahtalebi
Frank Rudzicz
151
12
0
13 Oct 2022
Assessing Neural Referential Form Selectors on a Realistic Multilingual
  Dataset
Assessing Neural Referential Form Selectors on a Realistic Multilingual Dataset
Guanyi Chen
F. Same
Kees van Deemter
111
0
0
10 Oct 2022
Survey: Exploiting Data Redundancy for Optimization of Deep Learning
Survey: Exploiting Data Redundancy for Optimization of Deep LearningACM Computing Surveys (ACM CSUR), 2022
Jou-An Chen
Wei Niu
Bin Ren
Yanzhi Wang
Xipeng Shen
136
28
0
29 Aug 2022
Proton: Probing Schema Linking Information from Pre-trained Language
  Models for Text-to-SQL Parsing
Proton: Probing Schema Linking Information from Pre-trained Language Models for Text-to-SQL ParsingKnowledge Discovery and Data Mining (KDD), 2022
Lihan Wang
Bowen Qin
Binyuan Hui
Bowen Li
Min Yang
Bailin Wang
Binhua Li
Fei Huang
Luo Si
Yongbin Li
175
49
0
28 Jun 2022
Analyzing Encoded Concepts in Transformer Language Models
Analyzing Encoded Concepts in Transformer Language ModelsNorth American Chapter of the Association for Computational Linguistics (NAACL), 2022
Hassan Sajjad
Nadir Durrani
Fahim Dalvi
Firoj Alam
A. Khan
Jia Xu
158
52
0
27 Jun 2022
Discovering Salient Neurons in Deep NLP Models
Discovering Salient Neurons in Deep NLP ModelsJournal of machine learning research (JMLR), 2022
Nadir Durrani
Fahim Dalvi
Hassan Sajjad
KELMMILM
263
19
0
27 Jun 2022
AST-Probe: Recovering abstract syntax trees from hidden representations
  of pre-trained language models
AST-Probe: Recovering abstract syntax trees from hidden representations of pre-trained language modelsInternational Conference on Automated Software Engineering (ASE), 2022
José Antonio Hernández López
Martin Weyssow
Jesús Sánchez Cuadrado
H. Sahraoui
139
27
0
23 Jun 2022
BridgeTower: Building Bridges Between Encoders in Vision-Language
  Representation Learning
BridgeTower: Building Bridges Between Encoders in Vision-Language Representation LearningAAAI Conference on Artificial Intelligence (AAAI), 2022
Xiao Xu
Chenfei Wu
Shachar Rosenman
Vasudev Lal
Wanxiang Che
Nan Duan
204
90
0
17 Jun 2022
Order-sensitive Shapley Values for Evaluating Conceptual Soundness of
  NLP Models
Order-sensitive Shapley Values for Evaluating Conceptual Soundness of NLP Models
Kaiji Lu
Anupam Datta
156
0
0
01 Jun 2022
Improving VAE-based Representation Learning
Improving VAE-based Representation Learning
Mingtian Zhang
Tim Z. Xiao
Brooks Paige
David Barber
SSLDRL
241
12
0
28 May 2022
DivEMT: Neural Machine Translation Post-Editing Effort Across
  Typologically Diverse Languages
DivEMT: Neural Machine Translation Post-Editing Effort Across Typologically Diverse LanguagesConference on Empirical Methods in Natural Language Processing (EMNLP), 2022
Gabriele Sarti
Arianna Bisazza
Ana Guerberof Arenas
Antonio Toral
208
11
0
24 May 2022
Discovering Latent Concepts Learned in BERT
Discovering Latent Concepts Learned in BERTInternational Conference on Learning Representations (ICLR), 2022
Fahim Dalvi
A. Khan
Firoj Alam
Nadir Durrani
Jia Xu
Hassan Sajjad
SSL
144
67
0
15 May 2022
Implicit N-grams Induced by Recurrence
Implicit N-grams Induced by RecurrenceNorth American Chapter of the Association for Computational Linguistics (NAACL), 2022
Xiaobing Sun
Wei Lu
148
4
0
05 May 2022
Systematicity, Compositionality and Transitivity of Deep NLP Models: a
  Metamorphic Testing Perspective
Systematicity, Compositionality and Transitivity of Deep NLP Models: a Metamorphic Testing PerspectiveFindings (Findings), 2022
Edoardo Manino
Julia Rozanova
Danilo S. Carvalho
André Freitas
Lucas C. Cordeiro
157
10
0
26 Apr 2022
It Takes Two Flints to Make a Fire: Multitask Learning of Neural
  Relation and Explanation Classifiers
It Takes Two Flints to Make a Fire: Multitask Learning of Neural Relation and Explanation ClassifiersInternational Conference on Computational Logic (ICCL), 2022
Zheng Tang
Mihai Surdeanu
411
9
0
25 Apr 2022
Probing Script Knowledge from Pre-Trained Models
Probing Script Knowledge from Pre-Trained Models
Zijian Jin
Xingyu Zhang
Mo Yu
Lifu Huang
151
5
0
16 Apr 2022
Interpretation of Black Box NLP Models: A Survey
Interpretation of Black Box NLP Models: A Survey
Shivani Choudhary
N. Chatterjee
S. K. Saha
FAtt
188
16
0
31 Mar 2022
Fine-Tuning can Distort Pretrained Features and Underperform
  Out-of-Distribution
Fine-Tuning can Distort Pretrained Features and Underperform Out-of-DistributionInternational Conference on Learning Representations (ICLR), 2022
Ananya Kumar
Aditi Raghunathan
Robbie Jones
Tengyu Ma
Abigail Z. Jacobs
OODD
292
813
0
21 Feb 2022
Evaluating the Construct Validity of Text Embeddings with Application to
  Survey Questions
Evaluating the Construct Validity of Text Embeddings with Application to Survey QuestionsEPJ Data Science (EPJ Data Sci.), 2022
Qixiang Fang
D. Nguyen
Daniel L. Oberski
192
15
0
18 Feb 2022
Locating and Editing Factual Associations in GPT
Locating and Editing Factual Associations in GPTNeural Information Processing Systems (NeurIPS), 2022
Kevin Meng
David Bau
A. Andonian
Yonatan Belinkov
KELM
871
1,886
0
10 Feb 2022
Table Pre-training: A Survey on Model Architectures, Pre-training
  Objectives, and Downstream Tasks
Table Pre-training: A Survey on Model Architectures, Pre-training Objectives, and Downstream TasksInternational Joint Conference on Artificial Intelligence (IJCAI), 2022
Haoyu Dong
Zhoujun Cheng
Xinyi He
Mengyuan Zhou
Anda Zhou
Fan Zhou
Ao Liu
Shi Han
Dongmei Zhang
LMTD
399
73
0
24 Jan 2022
Interpreting Arabic Transformer Models
Ahmed Abdelali
Nadir Durrani
Fahim Dalvi
Hassan Sajjad
136
2
0
19 Jan 2022
Representation Alignment in Neural Networks
Representation Alignment in Neural Networks
Ehsan Imani
Wei Hu
Martha White
169
6
0
15 Dec 2021
Variation and generality in encoding of syntactic anomaly information in
  sentence embeddings
Variation and generality in encoding of syntactic anomaly information in sentence embeddingsBlackboxNLP Workshop on Analyzing and Interpreting Neural Networks for NLP (BlackBoxNLP), 2021
Qinxuan Wu
Allyson Ettinger
157
2
0
12 Nov 2021
Fast Model Editing at Scale
Fast Model Editing at ScaleInternational Conference on Learning Representations (ICLR), 2021
E. Mitchell
Charles Lin
Antoine Bosselut
Chelsea Finn
Christopher D. Manning
KELM
873
451
0
21 Oct 2021
SlovakBERT: Slovak Masked Language Model
SlovakBERT: Slovak Masked Language Model
Matúš Pikuliak
Stefan Grivalsky
Martin Konopka
Miroslav Blšták
Martin Tamajka
Viktor Bachratý
Marian Simko
Pavol Balázik
Michal Trnka
Filip Uhlárik
166
33
0
30 Sep 2021
On the Prunability of Attention Heads in Multilingual BERT
On the Prunability of Attention Heads in Multilingual BERT
Aakriti Budhraja
Madhura Pande
Pratyush Kumar
Mitesh M. Khapra
156
5
0
26 Sep 2021
Previous
123456
Next