Communities
Connect sessions
AI calendar
Organizations
Join Slack
Contact Sales
Search
Open menu
Home
Papers
1504.00941
Cited By
v1
v2 (latest)
A Simple Way to Initialize Recurrent Networks of Rectified Linear Units
3 April 2015
Quoc V. Le
Navdeep Jaitly
Geoffrey E. Hinton
ODL
Re-assign community
ArXiv (abs)
PDF
HTML
Papers citing
"A Simple Way to Initialize Recurrent Networks of Rectified Linear Units"
50 / 353 papers shown
Solving hybrid machine learning tasks by traversing weight space geodesics
G. Raghavan
Matt Thomson
90
0
0
05 Jun 2021
Least Redundant Gated Recurrent Neural Network
IEEE International Joint Conference on Neural Network (IJCNN), 2021
Lukasz Neumann
Lukasz Lepak
Pawel Wawrzyñski
BDL
305
1
0
28 May 2021
Spectral Pruning for Recurrent Neural Networks
International Conference on Artificial Intelligence and Statistics (AISTATS), 2021
Takashi Furuya
Kazuma Suetake
K. Taniguchi
Hiroyuki Kusumoto
Ryuji Saiin
Tomohiro Daimon
179
4
0
23 May 2021
Slower is Better: Revisiting the Forgetting Mechanism in LSTM for Slower Information Decay
H. Chien
Javier S. Turek
Nicole M. Beckage
Vy A. Vo
C. Honey
Ted Willke
241
20
0
12 May 2021
Learning future terrorist targets through temporal meta-graphs
Scientific Reports (Sci Rep), 2021
G. Campedelli
Mihovil Bartulovic
Kathleen M. Carley
AI4TS
75
24
0
21 Apr 2021
Back to Square One: Superhuman Performance in Chutes and Ladders Through Deep Neural Networks and Tree Search
Dylan R. Ashley
Anssi Kanervisto
Brendan Bennett
254
1
0
01 Apr 2021
Continual Learning for Recurrent Neural Networks: an Empirical Evaluation
Neural Networks (NN), 2021
Andrea Cossu
Antonio Carta
Vincenzo Lomonaco
D. Bacciu
CLL
295
118
0
12 Mar 2021
Relational Weight Priors in Neural Networks for Abstract Pattern Learning and Language Modelling
R. Kopparti
Tillman Weyde
150
0
0
10 Mar 2021
UnICORNN: A recurrent model for learning very long time dependencies
International Conference on Machine Learning (ICML), 2021
T. Konstantin Rusch
Siddhartha Mishra
384
73
0
09 Mar 2021
An empirical analysis of phrase-based and neural machine translation
Hamidreza Ghader
113
1
0
04 Mar 2021
Parallelizing Legendre Memory Unit Training
International Conference on Machine Learning (ICML), 2021
Narsimha Chilkuri
C. Eliasmith
203
44
0
22 Feb 2021
Supervised training of spiking neural networks for robust deployment on mixed-signal neuromorphic processors
Scientific Reports (Sci Rep), 2021
Julian Büchel
D. Zendrikov
S. Solinas
Giacomo Indiveri
Dylan R. Muir
318
25
0
12 Feb 2021
Noisy Recurrent Neural Networks
Neural Information Processing Systems (NeurIPS), 2021
Soon Hoe Lim
N. Benjamin Erichson
Liam Hodgkinson
Michael W. Mahoney
295
66
0
09 Feb 2021
Ensemble perspective for understanding temporal credit assignment
Physical Review E (PRE), 2021
Wenxuan Zou
Chan Li
Haiping Huang
136
5
0
07 Feb 2021
CKConv: Continuous Kernel Convolution For Sequential Data
International Conference on Learning Representations (ICLR), 2021
David W. Romero
Anna Kuzina
Erik J. Bekkers
Jakub M. Tomczak
Mark Hoogendoorn
347
140
0
04 Feb 2021
S++: A Fast and Deployable Secure-Computation Framework for Privacy-Preserving Neural Network Training
Prashanthi Ramachandran
Shivam Agarwal
A. Mondal
Aastha Shah
Debayan Gupta
FedML
92
10
0
28 Jan 2021
Advances in Electron Microscopy with Deep Learning
Jeffrey M. Ede
666
3
0
04 Jan 2021
Kaleidoscope: An Efficient, Learnable Representation For All Structured Linear Maps
International Conference on Learning Representations (ICLR), 2020
Tri Dao
N. Sohoni
Albert Gu
Matthew Eichhorn
Amit Blonder
Megan Leszczynski
Atri Rudra
Christopher Ré
304
52
0
29 Dec 2020
Nanopore Base Calling on the Edge
Peter Perešíni
V. Boža
Broňa Brejová
T. Vinař
232
48
0
09 Nov 2020
FedSL: Federated Split Learning on Distributed Sequential Data in Recurrent Neural Networks
Ali Abedi
Shehroz S. Khan
FedML
248
80
0
06 Nov 2020
Short-Term Memory Optimization in Recurrent Neural Networks by Autoencoder-based Initialization
Antonio Carta
A. Sperduti
D. Bacciu
ODL
171
0
0
05 Nov 2020
Classification of Periodic Variable Stars with Novel Cyclic-Permutation Invariant Neural Networks
Monthly notices of the Royal Astronomical Society (MNRAS), 2020
Keming 名 Zhang 张 可
J. Bloom
243
12
0
02 Nov 2020
Recurrent Conditional Heteroskedasticity
Journal of applied econometrics (JAE), 2020
T.-N. Nguyen
Minh-Ngoc Tran
Robert Kohn
BDL
213
13
0
25 Oct 2020
CryptoGRU: Low Latency Privacy-Preserving Text Analysis With GRU
Bo Feng
Qian Lou
Lei Jiang
Geoffrey C. Fox
180
16
0
22 Oct 2020
RNN Training along Locally Optimal Trajectories via Frank-Wolfe Algorithm
Yun Yue
Ming Li
Venkatesh Saligrama
Ziming Zhang
324
5
0
12 Oct 2020
Revisiting Batch Normalization for Training Low-latency Deep Spiking Neural Networks from Scratch
Frontiers in Neuroscience (Front. Neurosci.), 2020
Youngeun Kim
Priyadarshini Panda
492
191
0
05 Oct 2020
Coupled Oscillatory Recurrent Neural Network (coRNN): An accurate and (gradient) stable architecture for learning long time dependencies
International Conference on Learning Representations (ICLR), 2020
T. Konstantin Rusch
Siddhartha Mishra
439
110
0
02 Oct 2020
Review: Deep Learning in Electron Microscopy
Jeffrey M. Ede
912
90
0
17 Sep 2020
Demystifying Deep Learning in Predictive Spatio-Temporal Analytics: An Information-Theoretic Framework
IEEE Transactions on Neural Networks and Learning Systems (IEEE TNNLS), 2020
Qi Tan
Yang Liu
Jiming Liu
AI4TS
231
9
0
14 Sep 2020
Online Spatiotemporal Action Detection and Prediction via Causal Representations
Gurkirt Singh
3DPC
CML
178
0
0
31 Aug 2020
HiPPO: Recurrent Memory with Optimal Polynomial Projections
Albert Gu
Tri Dao
Stefano Ermon
Atri Rudra
Christopher Ré
395
805
0
17 Aug 2020
Randomized Automatic Differentiation
Deniz Oktay
N. McGreivy
Joshua Aduol
Alex Beatson
Ryan P. Adams
ODL
160
28
0
20 Jul 2020
Shuffling Recurrent Neural Networks
AAAI Conference on Artificial Intelligence (AAAI), 2020
Michael Rotman
Lior Wolf
BDL
169
35
0
14 Jul 2020
Deep Isometric Learning for Visual Recognition
Haozhi Qi
Chong You
Xinyu Wang
Yi-An Ma
Jitendra Malik
VLM
191
56
0
30 Jun 2020
Lipschitz Recurrent Neural Networks
N. Benjamin Erichson
Omri Azencot
A. Queiruga
Liam Hodgkinson
Michael W. Mahoney
348
124
0
22 Jun 2020
The Recurrent Neural Tangent Kernel
Sina Alemohammad
Zichao Wang
Randall Balestriero
Richard Baraniuk
AAML
221
82
0
18 Jun 2020
Untangling tradeoffs between recurrence and self-attention in neural networks
Giancarlo Kerg
Bhargav Kanuparthi
Anirudh Goyal
Kyle Goyette
Yoshua Bengio
Guillaume Lajoie
170
9
0
16 Jun 2020
Artificial neural networks for neuroscientists: A primer
Neuron (Neuron), 2020
G. R. Yang
Xiao-Jing Wang
441
299
0
01 Jun 2020
Learning Various Length Dependence by Dual Recurrent Neural Networks
Neurocomputing (Neurocomputing), 2020
Chenpeng Zhang
Shuai Li
Mao Ye
Ce Zhu
Xue Li
123
9
0
28 May 2020
Effective and Efficient Computation with Multiple-timescale Spiking Recurrent Neural Networks
Bojian Yin
Federico Corradi
Sander M. Bohté
293
120
0
24 May 2020
Flexible Transmitter Network
Neural Computation (Neural Comput.), 2020
Shao-Qun Zhang
Zhi Zhou
216
14
0
08 Apr 2020
R-FORCE: Robust Learning for Random Recurrent Neural Networks
Yang Zheng
Eli Shlizerman
OOD
111
5
0
25 Mar 2020
Depth Enables Long-Term Memory for Recurrent Neural Networks
A. Ziv
92
0
0
23 Mar 2020
Interpretable Deep Recurrent Neural Networks via Unfolding Reweighted
ℓ
1
\ell_1
ℓ
1
-
ℓ
1
\ell_1
ℓ
1
Minimization: Architecture Design and Generalization Analysis
Huynh Van Luong
Boris Joukovsky
Nikos Deligiannis
166
4
0
18 Mar 2020
Refined Gate: A Simple and Effective Gating Mechanism for Recurrent Units
Zhanzhan Cheng
Yunlu Xu
Mingjian Cheng
Yu Qiao
Shiliang Pu
Yi Niu
Leilei Gan
111
10
0
26 Feb 2020
Evaluating complexity and resilience trade-offs in emerging memory inference machines
Neuro Inspired Computational Elements Workshop (NICE), 2020
C. Bennett
Ryan Dellana
T. Xiao
Ben Feinberg
S. Agarwal
S. Cardwell
M. Marinella
William M. Severa
Brad Aimone
114
2
0
25 Feb 2020
Scheduled Restart Momentum for Accelerated Stochastic Gradient Descent
SIAM Journal of Imaging Sciences (SIIMS), 2020
Bao Wang
T. Nguyen
Andrea L. Bertozzi
Richard G. Baraniuk
Stanley J. Osher
ODL
182
54
0
24 Feb 2020
A Spike in Performance: Training Hybrid-Spiking Neural Networks with Quantized Activation Functions
Aaron R. Voelker
Daniel Rasmussen
C. Eliasmith
190
18
0
10 Feb 2020
Encoding-based Memory Modules for Recurrent Neural Networks
Antonio Carta
A. Sperduti
D. Bacciu
KELM
109
2
0
31 Jan 2020
Provable Benefit of Orthogonal Initialization in Optimizing Deep Linear Networks
International Conference on Learning Representations (ICLR), 2020
Wei Hu
Lechao Xiao
Jeffrey Pennington
198
128
0
16 Jan 2020
Previous
1
2
3
4
5
6
7
8
Next