ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2026 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1504.00941
  4. Cited By
A Simple Way to Initialize Recurrent Networks of Rectified Linear Units
v1v2 (latest)

A Simple Way to Initialize Recurrent Networks of Rectified Linear Units

3 April 2015
Quoc V. Le
Navdeep Jaitly
Geoffrey E. Hinton
    ODL
ArXiv (abs)PDFHTML

Papers citing "A Simple Way to Initialize Recurrent Networks of Rectified Linear Units"

50 / 353 papers shown
Solving hybrid machine learning tasks by traversing weight space
  geodesics
Solving hybrid machine learning tasks by traversing weight space geodesics
G. Raghavan
Matt Thomson
90
0
0
05 Jun 2021
Least Redundant Gated Recurrent Neural Network
Least Redundant Gated Recurrent Neural NetworkIEEE International Joint Conference on Neural Network (IJCNN), 2021
Lukasz Neumann
Lukasz Lepak
Pawel Wawrzyñski
BDL
305
1
0
28 May 2021
Spectral Pruning for Recurrent Neural Networks
Spectral Pruning for Recurrent Neural NetworksInternational Conference on Artificial Intelligence and Statistics (AISTATS), 2021
Takashi Furuya
Kazuma Suetake
K. Taniguchi
Hiroyuki Kusumoto
Ryuji Saiin
Tomohiro Daimon
179
4
0
23 May 2021
Slower is Better: Revisiting the Forgetting Mechanism in LSTM for Slower
  Information Decay
Slower is Better: Revisiting the Forgetting Mechanism in LSTM for Slower Information Decay
H. Chien
Javier S. Turek
Nicole M. Beckage
Vy A. Vo
C. Honey
Ted Willke
241
20
0
12 May 2021
Learning future terrorist targets through temporal meta-graphs
Learning future terrorist targets through temporal meta-graphsScientific Reports (Sci Rep), 2021
G. Campedelli
Mihovil Bartulovic
Kathleen M. Carley
AI4TS
75
24
0
21 Apr 2021
Back to Square One: Superhuman Performance in Chutes and Ladders Through
  Deep Neural Networks and Tree Search
Back to Square One: Superhuman Performance in Chutes and Ladders Through Deep Neural Networks and Tree Search
Dylan R. Ashley
Anssi Kanervisto
Brendan Bennett
254
1
0
01 Apr 2021
Continual Learning for Recurrent Neural Networks: an Empirical
  Evaluation
Continual Learning for Recurrent Neural Networks: an Empirical EvaluationNeural Networks (NN), 2021
Andrea Cossu
Antonio Carta
Vincenzo Lomonaco
D. Bacciu
CLL
295
118
0
12 Mar 2021
Relational Weight Priors in Neural Networks for Abstract Pattern
  Learning and Language Modelling
Relational Weight Priors in Neural Networks for Abstract Pattern Learning and Language Modelling
R. Kopparti
Tillman Weyde
150
0
0
10 Mar 2021
UnICORNN: A recurrent model for learning very long time dependencies
UnICORNN: A recurrent model for learning very long time dependenciesInternational Conference on Machine Learning (ICML), 2021
T. Konstantin Rusch
Siddhartha Mishra
384
73
0
09 Mar 2021
An empirical analysis of phrase-based and neural machine translation
An empirical analysis of phrase-based and neural machine translation
Hamidreza Ghader
113
1
0
04 Mar 2021
Parallelizing Legendre Memory Unit Training
Parallelizing Legendre Memory Unit TrainingInternational Conference on Machine Learning (ICML), 2021
Narsimha Chilkuri
C. Eliasmith
203
44
0
22 Feb 2021
Supervised training of spiking neural networks for robust deployment on
  mixed-signal neuromorphic processors
Supervised training of spiking neural networks for robust deployment on mixed-signal neuromorphic processorsScientific Reports (Sci Rep), 2021
Julian Büchel
D. Zendrikov
S. Solinas
Giacomo Indiveri
Dylan R. Muir
318
25
0
12 Feb 2021
Noisy Recurrent Neural Networks
Noisy Recurrent Neural NetworksNeural Information Processing Systems (NeurIPS), 2021
Soon Hoe Lim
N. Benjamin Erichson
Liam Hodgkinson
Michael W. Mahoney
295
66
0
09 Feb 2021
Ensemble perspective for understanding temporal credit assignment
Ensemble perspective for understanding temporal credit assignmentPhysical Review E (PRE), 2021
Wenxuan Zou
Chan Li
Haiping Huang
136
5
0
07 Feb 2021
CKConv: Continuous Kernel Convolution For Sequential Data
CKConv: Continuous Kernel Convolution For Sequential DataInternational Conference on Learning Representations (ICLR), 2021
David W. Romero
Anna Kuzina
Erik J. Bekkers
Jakub M. Tomczak
Mark Hoogendoorn
347
140
0
04 Feb 2021
S++: A Fast and Deployable Secure-Computation Framework for
  Privacy-Preserving Neural Network Training
S++: A Fast and Deployable Secure-Computation Framework for Privacy-Preserving Neural Network Training
Prashanthi Ramachandran
Shivam Agarwal
A. Mondal
Aastha Shah
Debayan Gupta
FedML
92
10
0
28 Jan 2021
Advances in Electron Microscopy with Deep Learning
Advances in Electron Microscopy with Deep Learning
Jeffrey M. Ede
666
3
0
04 Jan 2021
Kaleidoscope: An Efficient, Learnable Representation For All Structured
  Linear Maps
Kaleidoscope: An Efficient, Learnable Representation For All Structured Linear MapsInternational Conference on Learning Representations (ICLR), 2020
Tri Dao
N. Sohoni
Albert Gu
Matthew Eichhorn
Amit Blonder
Megan Leszczynski
Atri Rudra
Christopher Ré
304
52
0
29 Dec 2020
Nanopore Base Calling on the Edge
Nanopore Base Calling on the Edge
Peter Perešíni
V. Boža
Broňa Brejová
T. Vinař
232
48
0
09 Nov 2020
FedSL: Federated Split Learning on Distributed Sequential Data in
  Recurrent Neural Networks
FedSL: Federated Split Learning on Distributed Sequential Data in Recurrent Neural Networks
Ali Abedi
Shehroz S. Khan
FedML
248
80
0
06 Nov 2020
Short-Term Memory Optimization in Recurrent Neural Networks by
  Autoencoder-based Initialization
Short-Term Memory Optimization in Recurrent Neural Networks by Autoencoder-based Initialization
Antonio Carta
A. Sperduti
D. Bacciu
ODL
171
0
0
05 Nov 2020
Classification of Periodic Variable Stars with Novel Cyclic-Permutation
  Invariant Neural Networks
Classification of Periodic Variable Stars with Novel Cyclic-Permutation Invariant Neural NetworksMonthly notices of the Royal Astronomical Society (MNRAS), 2020
Keming 名 Zhang 张 可
J. Bloom
243
12
0
02 Nov 2020
Recurrent Conditional Heteroskedasticity
Recurrent Conditional HeteroskedasticityJournal of applied econometrics (JAE), 2020
T.-N. Nguyen
Minh-Ngoc Tran
Robert Kohn
BDL
213
13
0
25 Oct 2020
CryptoGRU: Low Latency Privacy-Preserving Text Analysis With GRU
CryptoGRU: Low Latency Privacy-Preserving Text Analysis With GRU
Bo Feng
Qian Lou
Lei Jiang
Geoffrey C. Fox
180
16
0
22 Oct 2020
RNN Training along Locally Optimal Trajectories via Frank-Wolfe
  Algorithm
RNN Training along Locally Optimal Trajectories via Frank-Wolfe Algorithm
Yun Yue
Ming Li
Venkatesh Saligrama
Ziming Zhang
324
5
0
12 Oct 2020
Revisiting Batch Normalization for Training Low-latency Deep Spiking
  Neural Networks from Scratch
Revisiting Batch Normalization for Training Low-latency Deep Spiking Neural Networks from ScratchFrontiers in Neuroscience (Front. Neurosci.), 2020
Youngeun Kim
Priyadarshini Panda
492
191
0
05 Oct 2020
Coupled Oscillatory Recurrent Neural Network (coRNN): An accurate and
  (gradient) stable architecture for learning long time dependencies
Coupled Oscillatory Recurrent Neural Network (coRNN): An accurate and (gradient) stable architecture for learning long time dependenciesInternational Conference on Learning Representations (ICLR), 2020
T. Konstantin Rusch
Siddhartha Mishra
439
110
0
02 Oct 2020
Review: Deep Learning in Electron Microscopy
Review: Deep Learning in Electron Microscopy
Jeffrey M. Ede
912
90
0
17 Sep 2020
Demystifying Deep Learning in Predictive Spatio-Temporal Analytics: An
  Information-Theoretic Framework
Demystifying Deep Learning in Predictive Spatio-Temporal Analytics: An Information-Theoretic FrameworkIEEE Transactions on Neural Networks and Learning Systems (IEEE TNNLS), 2020
Qi Tan
Yang Liu
Jiming Liu
AI4TS
231
9
0
14 Sep 2020
Online Spatiotemporal Action Detection and Prediction via Causal
  Representations
Online Spatiotemporal Action Detection and Prediction via Causal Representations
Gurkirt Singh
3DPCCML
178
0
0
31 Aug 2020
HiPPO: Recurrent Memory with Optimal Polynomial Projections
HiPPO: Recurrent Memory with Optimal Polynomial Projections
Albert Gu
Tri Dao
Stefano Ermon
Atri Rudra
Christopher Ré
395
805
0
17 Aug 2020
Randomized Automatic Differentiation
Randomized Automatic Differentiation
Deniz Oktay
N. McGreivy
Joshua Aduol
Alex Beatson
Ryan P. Adams
ODL
160
28
0
20 Jul 2020
Shuffling Recurrent Neural Networks
Shuffling Recurrent Neural NetworksAAAI Conference on Artificial Intelligence (AAAI), 2020
Michael Rotman
Lior Wolf
BDL
169
35
0
14 Jul 2020
Deep Isometric Learning for Visual Recognition
Deep Isometric Learning for Visual Recognition
Haozhi Qi
Chong You
Xinyu Wang
Yi-An Ma
Jitendra Malik
VLM
191
56
0
30 Jun 2020
Lipschitz Recurrent Neural Networks
Lipschitz Recurrent Neural Networks
N. Benjamin Erichson
Omri Azencot
A. Queiruga
Liam Hodgkinson
Michael W. Mahoney
348
124
0
22 Jun 2020
The Recurrent Neural Tangent Kernel
The Recurrent Neural Tangent Kernel
Sina Alemohammad
Zichao Wang
Randall Balestriero
Richard Baraniuk
AAML
221
82
0
18 Jun 2020
Untangling tradeoffs between recurrence and self-attention in neural
  networks
Untangling tradeoffs between recurrence and self-attention in neural networks
Giancarlo Kerg
Bhargav Kanuparthi
Anirudh Goyal
Kyle Goyette
Yoshua Bengio
Guillaume Lajoie
170
9
0
16 Jun 2020
Artificial neural networks for neuroscientists: A primer
Artificial neural networks for neuroscientists: A primerNeuron (Neuron), 2020
G. R. Yang
Xiao-Jing Wang
441
299
0
01 Jun 2020
Learning Various Length Dependence by Dual Recurrent Neural Networks
Learning Various Length Dependence by Dual Recurrent Neural NetworksNeurocomputing (Neurocomputing), 2020
Chenpeng Zhang
Shuai Li
Mao Ye
Ce Zhu
Xue Li
123
9
0
28 May 2020
Effective and Efficient Computation with Multiple-timescale Spiking
  Recurrent Neural Networks
Effective and Efficient Computation with Multiple-timescale Spiking Recurrent Neural Networks
Bojian Yin
Federico Corradi
Sander M. Bohté
293
120
0
24 May 2020
Flexible Transmitter Network
Flexible Transmitter NetworkNeural Computation (Neural Comput.), 2020
Shao-Qun Zhang
Zhi Zhou
216
14
0
08 Apr 2020
R-FORCE: Robust Learning for Random Recurrent Neural Networks
R-FORCE: Robust Learning for Random Recurrent Neural Networks
Yang Zheng
Eli Shlizerman
OOD
111
5
0
25 Mar 2020
Depth Enables Long-Term Memory for Recurrent Neural Networks
Depth Enables Long-Term Memory for Recurrent Neural Networks
A. Ziv
92
0
0
23 Mar 2020
Interpretable Deep Recurrent Neural Networks via Unfolding Reweighted
  $\ell_1$-$\ell_1$ Minimization: Architecture Design and Generalization
  Analysis
Interpretable Deep Recurrent Neural Networks via Unfolding Reweighted ℓ1\ell_1ℓ1​-ℓ1\ell_1ℓ1​ Minimization: Architecture Design and Generalization Analysis
Huynh Van Luong
Boris Joukovsky
Nikos Deligiannis
166
4
0
18 Mar 2020
Refined Gate: A Simple and Effective Gating Mechanism for Recurrent
  Units
Refined Gate: A Simple and Effective Gating Mechanism for Recurrent Units
Zhanzhan Cheng
Yunlu Xu
Mingjian Cheng
Yu Qiao
Shiliang Pu
Yi Niu
Leilei Gan
111
10
0
26 Feb 2020
Evaluating complexity and resilience trade-offs in emerging memory
  inference machines
Evaluating complexity and resilience trade-offs in emerging memory inference machinesNeuro Inspired Computational Elements Workshop (NICE), 2020
C. Bennett
Ryan Dellana
T. Xiao
Ben Feinberg
S. Agarwal
S. Cardwell
M. Marinella
William M. Severa
Brad Aimone
114
2
0
25 Feb 2020
Scheduled Restart Momentum for Accelerated Stochastic Gradient Descent
Scheduled Restart Momentum for Accelerated Stochastic Gradient DescentSIAM Journal of Imaging Sciences (SIIMS), 2020
Bao Wang
T. Nguyen
Andrea L. Bertozzi
Richard G. Baraniuk
Stanley J. Osher
ODL
182
54
0
24 Feb 2020
A Spike in Performance: Training Hybrid-Spiking Neural Networks with
  Quantized Activation Functions
A Spike in Performance: Training Hybrid-Spiking Neural Networks with Quantized Activation Functions
Aaron R. Voelker
Daniel Rasmussen
C. Eliasmith
190
18
0
10 Feb 2020
Encoding-based Memory Modules for Recurrent Neural Networks
Encoding-based Memory Modules for Recurrent Neural Networks
Antonio Carta
A. Sperduti
D. Bacciu
KELM
109
2
0
31 Jan 2020
Provable Benefit of Orthogonal Initialization in Optimizing Deep Linear
  Networks
Provable Benefit of Orthogonal Initialization in Optimizing Deep Linear NetworksInternational Conference on Learning Representations (ICLR), 2020
Wei Hu
Lechao Xiao
Jeffrey Pennington
198
128
0
16 Jan 2020
Previous
12345678
Next