ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1805.04218
  4. Cited By
Deep RNNs Encode Soft Hierarchical Syntax

Deep RNNs Encode Soft Hierarchical Syntax

11 May 2018
Terra Blevins
Omer Levy
Luke Zettlemoyer
ArXivPDFHTML

Papers citing "Deep RNNs Encode Soft Hierarchical Syntax"

28 / 78 papers shown
Title
The Bottom-up Evolution of Representations in the Transformer: A Study
  with Machine Translation and Language Modeling Objectives
The Bottom-up Evolution of Representations in the Transformer: A Study with Machine Translation and Language Modeling Objectives
Elena Voita
Rico Sennrich
Ivan Titov
201
181
0
03 Sep 2019
Higher-order Comparisons of Sentence Encoder Representations
Higher-order Comparisons of Sentence Encoder Representations
Mostafa Abdou
Artur Kulmizev
Felix Hill
D. Low
Anders Søgaard
22
16
0
01 Sep 2019
Shallow Syntax in Deep Water
Shallow Syntax in Deep Water
Swabha Swayamdipta
Matthew E. Peters
Brendan Roof
Chris Dyer
Noah A. Smith
16
10
0
29 Aug 2019
Multiresolution Transformer Networks: Recurrence is Not Essential for
  Modeling Hierarchical Structure
Multiresolution Transformer Networks: Recurrence is Not Essential for Modeling Hierarchical Structure
Vikas K. Garg
Inderjit S. Dhillon
Hsiang-Fu Yu
18
7
0
27 Aug 2019
Compositionality decomposed: how do neural networks generalise?
Compositionality decomposed: how do neural networks generalise?
Dieuwke Hupkes
Verna Dankers
Mathijs Mul
Elia Bruni
CoGe
28
321
0
22 Aug 2019
Cross-Lingual Transfer for Distantly Supervised and Low-resources
  Indonesian NER
Cross-Lingual Transfer for Distantly Supervised and Low-resources Indonesian NER
Fariz Ikhwantri
6
3
0
25 Jul 2019
What Should/Do/Can LSTMs Learn When Parsing Auxiliary Verb
  Constructions?
What Should/Do/Can LSTMs Learn When Parsing Auxiliary Verb Constructions?
Miryam de Lhoneux
Sara Stymne
Joakim Nivre
12
3
0
18 Jul 2019
A Comparative Analysis of Knowledge-Intensive and Data-Intensive
  Semantic Parsers
A Comparative Analysis of Knowledge-Intensive and Data-Intensive Semantic Parsers
Junjie Cao
Zi-yu Lin
Weiwei SUN
Xiaojun Wan
13
1
0
04 Jul 2019
Scalable Syntax-Aware Language Models Using Knowledge Distillation
Scalable Syntax-Aware Language Models Using Knowledge Distillation
A. Kuncoro
Chris Dyer
Laura Rimell
S. Clark
Phil Blunsom
35
26
0
14 Jun 2019
What Does BERT Look At? An Analysis of BERT's Attention
What Does BERT Look At? An Analysis of BERT's Attention
Kevin Clark
Urvashi Khandelwal
Omer Levy
Christopher D. Manning
MILM
78
1,581
0
11 Jun 2019
Analyzing the Structure of Attention in a Transformer Language Model
Analyzing the Structure of Attention in a Transformer Language Model
Jesse Vig
Yonatan Belinkov
30
357
0
07 Jun 2019
Visualizing and Measuring the Geometry of BERT
Visualizing and Measuring the Geometry of BERT
Andy Coenen
Emily Reif
Ann Yuan
Been Kim
Adam Pearce
F. Viégas
Martin Wattenberg
MILM
43
415
0
06 Jun 2019
From Balustrades to Pierre Vinken: Looking for Syntax in Transformer
  Self-Attentions
From Balustrades to Pierre Vinken: Looking for Syntax in Transformer Self-Attentions
David Marecek
Rudolf Rosa
8
52
0
05 Jun 2019
Assessing the Ability of Self-Attention Networks to Learn Word Order
Assessing the Ability of Self-Attention Networks to Learn Word Order
Baosong Yang
Longyue Wang
Derek F. Wong
Lidia S. Chao
Zhaopeng Tu
11
31
0
03 Jun 2019
The (Non-)Utility of Structural Features in BiLSTM-based Dependency
  Parsers
The (Non-)Utility of Structural Features in BiLSTM-based Dependency Parsers
Agnieszka Falenska
Jonas Kuhn
9
19
0
29 May 2019
What do you learn from context? Probing for sentence structure in
  contextualized word representations
What do you learn from context? Probing for sentence structure in contextualized word representations
Ian Tenney
Patrick Xia
Berlin Chen
Alex Jinpeng Wang
Adam Poliak
...
Najoung Kim
Benjamin Van Durme
Samuel R. Bowman
Dipanjan Das
Ellie Pavlick
91
848
0
15 May 2019
BERT Rediscovers the Classical NLP Pipeline
BERT Rediscovers the Classical NLP Pipeline
Ian Tenney
Dipanjan Das
Ellie Pavlick
MILM
SSeg
35
1,438
0
15 May 2019
Probing Biomedical Embeddings from Language Models
Probing Biomedical Embeddings from Language Models
Qiao Jin
Bhuwan Dhingra
William W. Cohen
Xinghua Lu
24
116
0
03 Apr 2019
Modeling Vocabulary for Big Code Machine Learning
Modeling Vocabulary for Big Code Machine Learning
Hlib Babii
Andrea Janes
Romain Robbes
11
22
0
03 Apr 2019
Linguistic Knowledge and Transferability of Contextual Representations
Linguistic Knowledge and Transferability of Contextual Representations
Nelson F. Liu
Matt Gardner
Yonatan Belinkov
Matthew E. Peters
Noah A. Smith
52
717
0
21 Mar 2019
Pretrained language model transfer on neural named entity recognition in
  Indonesian conversational texts
Pretrained language model transfer on neural named entity recognition in Indonesian conversational texts
Rezka Leonandya
Fariz Ikhwantri
14
4
0
21 Feb 2019
Analysis Methods in Neural Language Processing: A Survey
Analysis Methods in Neural Language Processing: A Survey
Yonatan Belinkov
James R. Glass
39
547
0
21 Dec 2018
Understanding Learning Dynamics Of Language Models with SVCCA
Understanding Learning Dynamics Of Language Models with SVCCA
Naomi Saphra
Adam Lopez
8
94
0
01 Nov 2018
RNNs as psycholinguistic subjects: Syntactic state and grammatical
  dependency
RNNs as psycholinguistic subjects: Syntactic state and grammatical dependency
Richard Futrell
Ethan Gotlieb Wilcox
Takashi Morita
R. Levy
11
57
0
05 Sep 2018
On Tree-Based Neural Sentence Modeling
On Tree-Based Neural Sentence Modeling
Freda Shi
Hao Zhou
Jiaze Chen
Lei Li
AI4CE
27
42
0
29 Aug 2018
Dissecting Contextual Word Embeddings: Architecture and Representation
Dissecting Contextual Word Embeddings: Architecture and Representation
Matthew E. Peters
Mark Neumann
Luke Zettlemoyer
Wen-tau Yih
33
425
0
27 Aug 2018
The Importance of Being Recurrent for Modeling Hierarchical Structure
The Importance of Being Recurrent for Modeling Hierarchical Structure
Ke M. Tran
Arianna Bisazza
Christof Monz
32
150
0
09 Mar 2018
OpenNMT: Open-Source Toolkit for Neural Machine Translation
OpenNMT: Open-Source Toolkit for Neural Machine Translation
Guillaume Klein
Yoon Kim
Yuntian Deng
Jean Senellart
Alexander M. Rush
271
1,896
0
10 Jan 2017
Previous
12