ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2026 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1606.01305
  4. Cited By
Zoneout: Regularizing RNNs by Randomly Preserving Hidden Activations
v1v2v3v4 (latest)

Zoneout: Regularizing RNNs by Randomly Preserving Hidden Activations

International Conference on Learning Representations (ICLR), 2016
3 June 2016
David M. Krueger
Tegan Maharaj
János Kramár
Mohammad Pezeshki
Nicolas Ballas
Nan Rosemary Ke
Anirudh Goyal
Yoshua Bengio
Aaron Courville
C. Pal
ArXiv (abs)PDFHTML

Papers citing "Zoneout: Regularizing RNNs by Randomly Preserving Hidden Activations"

50 / 180 papers shown
Stroke-based sketched symbol reconstruction and segmentation
Stroke-based sketched symbol reconstruction and segmentation
Kurmanbek Kaiyrbekov
M. Sezgin
184
20
0
10 Jan 2019
Learning latent representations for style control and transfer in
  end-to-end speech synthesis
Learning latent representations for style control and transfer in end-to-end speech synthesis
Ya-Jie Zhang
Shifeng Pan
Lei He
Zhenhua Ling
BDLSSLDRL
193
240
0
11 Dec 2018
Drop-Activation: Implicit Parameter Reduction and Harmonic
  Regularization
Drop-Activation: Implicit Parameter Reduction and Harmonic Regularization
Senwei Liang
Y. Khoo
Haizhao Yang
AI4CE
150
2
0
14 Nov 2018
Benchmarking Deep Sequential Models on Volatility Predictions for
  Financial Time Series
Benchmarking Deep Sequential Models on Volatility Predictions for Financial Time Series
Qiang Zhang
Kyle Birkeland
Yaodong Yang
Yixiao Liu
138
9
0
08 Nov 2018
Cycle-consistency training for end-to-end speech recognition
Cycle-consistency training for end-to-end speech recognition
Takaaki Hori
Ramón Fernández Astudillo
Tomoki Hayashi
Yu Zhang
Shinji Watanabe
Jonathan Le Roux
179
89
0
02 Nov 2018
DropBlock: A regularization method for convolutional networks
DropBlock: A regularization method for convolutional networks
Golnaz Ghiasi
Nayeon Lee
Quoc V. Le
272
1,018
0
30 Oct 2018
Investigation of enhanced Tacotron text-to-speech synthesis systems with
  self-attention for pitch accent language
Investigation of enhanced Tacotron text-to-speech synthesis systems with self-attention for pitch accent language
Yusuke Yasuda
Xin Wang
Shinji Takaki
Junichi Yamagishi
197
89
0
29 Oct 2018
Sequence-to-Sequence Acoustic Modeling for Voice Conversion
Sequence-to-Sequence Acoustic Modeling for Voice Conversion
Jing-Xuan Zhang
Zhenhua Ling
Li-Juan Liu
Yuan Jiang
Lirong Dai
356
136
0
16 Oct 2018
Dropout as a Structured Shrinkage Prior
Dropout as a Structured Shrinkage Prior
Eric T. Nalisnick
José Miguel Hernández-Lobato
Padhraic Smyth
BDLUQCV
217
1
0
09 Oct 2018
h-detach: Modifying the LSTM Gradient Towards Better Optimization
h-detach: Modifying the LSTM Gradient Towards Better Optimization
Devansh Arpit
Bhargav Kanuparthi
Giancarlo Kerg
Nan Rosemary Ke
Alexia Jolicoeur-Martineau
Yoshua Bengio
312
36
0
06 Oct 2018
Semi-Supervised Training for Improving Data Efficiency in End-to-End
  Speech Synthesis
Semi-Supervised Training for Improving Data Efficiency in End-to-End Speech Synthesis
Yu-An Chung
Yuxuan Wang
Wei-Ning Hsu
Yu Zhang
RJ Skerry-Ryan
138
120
0
30 Aug 2018
Dropout with Tabu Strategy for Regularizing Deep Neural Networks
Dropout with Tabu Strategy for Regularizing Deep Neural Networks
Zongjie Ma
A. Sattar
Jun Zhou
Qingliang Chen
Kaile Su
159
6
0
29 Aug 2018
Neural Architecture Optimization
Neural Architecture Optimization
Renqian Luo
Fei Tian
Tao Qin
Enhong Chen
Tie-Yan Liu
3DV
439
693
0
22 Aug 2018
Improved Language Modeling by Decoding the Past
Improved Language Modeling by Decoding the Past
Siddhartha Brahma
BDLAI4TS
275
6
0
14 Aug 2018
Confidence penalty, annealing Gaussian noise and zoneout for biLSTM-CRF
  networks for named entity recognition
Confidence penalty, annealing Gaussian noise and zoneout for biLSTM-CRF networks for named entity recognition
Antonio Jimeno Yepes
112
2
0
13 Aug 2018
Character-Level Language Modeling with Deeper Self-Attention
Character-Level Language Modeling with Deeper Self-Attention
Rami Al-Rfou
Dokook Choe
Noah Constant
Mandy Guo
Llion Jones
346
409
0
09 Aug 2018
Back-Translation-Style Data Augmentation for End-to-End ASR
Back-Translation-Style Data Augmentation for End-to-End ASR
Tomoki Hayashi
Shinji Watanabe
Yu Zhang
Tomoki Toda
Takaaki Hori
Ramón Fernández Astudillo
K. Takeda
174
109
0
28 Jul 2018
Effectiveness of Scaled Exponentially-Regularized Linear Units (SERLUs)
Effectiveness of Scaled Exponentially-Regularized Linear Units (SERLUs)
Guoqiang Zhang
Hao Li
112
11
0
26 Jul 2018
Recent Advances in Deep Learning: An Overview
Recent Advances in Deep Learning: An Overview
Matiur Rahman Minar
Jibon Naher
VLM
188
131
0
21 Jul 2018
Recurrent DNNs and its Ensembles on the TIMIT Phone Recognition Task
Recurrent DNNs and its Ensembles on the TIMIT Phone Recognition Task
Jan Vaněk
Josef Michálek
J. Psutka
UQCV
77
5
0
19 Jun 2018
Towards Binary-Valued Gates for Robust LSTM Training
Towards Binary-Valued Gates for Robust LSTM Training
Zhuohan Li
Di He
Fei Tian
Wei-neng Chen
Tao Qin
Liwei Wang
Tie-Yan Liu
MQ
149
49
0
08 Jun 2018
Efficient Full-Matrix Adaptive Regularization
Efficient Full-Matrix Adaptive Regularization
Naman Agarwal
Brian Bullins
Xinyi Chen
Elad Hazan
Karan Singh
Cyril Zhang
Yi Zhang
136
22
0
08 Jun 2018
Grow and Prune Compact, Fast, and Accurate LSTMs
Grow and Prune Compact, Fast, and Accurate LSTMs
Xiaoliang Dai
Hongxu Yin
N. Jha
VLMSyDa
181
98
0
30 May 2018
Highway State Gating for Recurrent Highway Networks: improving
  information flow through time
Highway State Gating for Recurrent Highway Networks: improving information flow through time
Ron Shoham
Haim Permuter
67
0
0
23 May 2018
Token-level and sequence-level loss smoothing for RNN language models
Token-level and sequence-level loss smoothing for RNN language models
Maha Elbayad
Laurent Besacier
Jakob Verbeek
154
19
0
14 May 2018
Noisin: Unbiased Regularization for Recurrent Neural Networks
Noisin: Unbiased Regularization for Recurrent Neural Networks
Adji Bousso Dieng
Rajesh Ranganath
Jaan Altosaar
David M. Blei
145
25
0
03 May 2018
The unreasonable effectiveness of the forget gate
The unreasonable effectiveness of the forget gate
J. Westhuizen
Joan Lasenby
180
98
0
13 Apr 2018
Regularizing RNNs for Caption Generation by Reconstructing The Past with
  The Present
Regularizing RNNs for Caption Generation by Reconstructing The Past with The Present
Xinpeng Chen
Lin Ma
Wenhao Jiang
Jian Yao
Wen Liu
215
95
0
30 Mar 2018
Style Tokens: Unsupervised Style Modeling, Control and Transfer in
  End-to-End Speech Synthesis
Style Tokens: Unsupervised Style Modeling, Control and Transfer in End-to-End Speech Synthesis
Yuxuan Wang
Daisy Stanton
Yu Zhang
RJ Skerry-Ryan
Eric Battenberg
Joel Shor
Y. Xiao
Fei Ren
Ye Jia
Rif A. Saurous
286
890
0
23 Mar 2018
An Analysis of Neural Language Modeling at Multiple Scales
An Analysis of Neural Language Modeling at Multiple Scales
Stephen Merity
N. Keskar
R. Socher
180
172
0
22 Mar 2018
Independently Recurrent Neural Network (IndRNN): Building A Longer and
  Deeper RNN
Independently Recurrent Neural Network (IndRNN): Building A Longer and Deeper RNN
Shuai Li
W. Li
Chris Cook
Ce Zhu
Yanbo Gao
369
797
0
13 Mar 2018
Flipout: Efficient Pseudo-Independent Weight Perturbations on
  Mini-Batches
Flipout: Efficient Pseudo-Independent Weight Perturbations on Mini-BatchesInternational Conference on Learning Representations (ICLR), 2018
Yeming Wen
Paul Vicol
Jimmy Ba
Dustin Tran
Roger C. Grosse
BDL
413
338
0
12 Mar 2018
An Empirical Evaluation of Generic Convolutional and Recurrent Networks
  for Sequence Modeling
An Empirical Evaluation of Generic Convolutional and Recurrent Networks for Sequence Modeling
Shaojie Bai
J. Zico Kolter
V. Koltun
DRL
360
5,922
0
04 Mar 2018
Nested LSTMs
Nested LSTMs
Joel Ruben Antony Moniz
David M. Krueger
175
70
0
31 Jan 2018
Scalable and accurate deep learning for electronic health records
Scalable and accurate deep learning for electronic health records
A. Rajkomar
Eyal Oren
Kai Chen
Andrew M. Dai
Nissan Hajaj
...
A. Butte
M. Howell
Claire Cui
Greg S. Corrado
Jeffrey Dean
OODBDL
426
2,441
0
24 Jan 2018
Recent Advances in Recurrent Neural Networks
Recent Advances in Recurrent Neural Networks
Hojjat Salehinejad
Sharan Sankar
Joseph Barfett
E. Colak
S. Valaee
AI4TS
499
692
0
29 Dec 2017
Natural TTS Synthesis by Conditioning WaveNet on Mel Spectrogram
  Predictions
Natural TTS Synthesis by Conditioning WaveNet on Mel Spectrogram Predictions
Jonathan Shen
Ruoming Pang
Ron J. Weiss
M. Schuster
Navdeep Jaitly
...
Yuxuan Wang
RJ Skerry-Ryan
Rif A. Saurous
Yannis Agiomyrgiannakis
Yonghui Wu
450
2,937
0
16 Dec 2017
Variational Bi-LSTMs
Variational Bi-LSTMs
Samira Shabanian
Devansh Arpit
Adam Trischler
Yoshua Bengio
DRL
114
28
0
15 Nov 2017
Fine-tuning Tree-LSTM for phrase-level sentiment classification on a
  Polish dependency treebank. Submission to PolEval task 2
Fine-tuning Tree-LSTM for phrase-level sentiment classification on a Polish dependency treebank. Submission to PolEval task 2
Tomasz Korbak
Paulina Zak
53
3
0
03 Nov 2017
Neural Language Modeling by Jointly Learning Syntax and Lexicon
Neural Language Modeling by Jointly Learning Syntax and Lexicon
Songlin Yang
Zhouhan Lin
Chin-Wei Huang
Aaron Courville
224
182
0
02 Nov 2017
Fraternal Dropout
Fraternal Dropout
Konrad Zolna
Devansh Arpit
Dendi Suhubdy
Yoshua Bengio
186
53
0
31 Oct 2017
Rotational Unit of Memory
Rotational Unit of MemoryInternational Conference on Learning Representations (ICLR), 2017
Rumen Dangovski
L. Jing
Marin Soljacic
163
7
0
26 Oct 2017
Dilated Recurrent Neural Networks
Dilated Recurrent Neural Networks
Shiyu Chang
Yang Zhang
Wei Han
Mo Yu
Xiaoxiao Guo
Wei Tan
Xiaodong Cui
Michael Witbrock
M. Hasegawa-Johnson
Thomas S. Huang
312
338
0
05 Oct 2017
Shifting Mean Activation Towards Zero with Bipolar Activation Functions
Shifting Mean Activation Towards Zero with Bipolar Activation Functions
L. Eidnes
Arild Nøkland
200
18
0
12 Sep 2017
Skip RNN: Learning to Skip State Updates in Recurrent Neural Networks
Skip RNN: Learning to Skip State Updates in Recurrent Neural Networks
Victor Campos
Brendan Jou
Xavier Giró-i-Nieto
Jordi Torres
Shih-Fu Chang
301
227
0
22 Aug 2017
Twin Networks: Matching the Future for Sequence Generation
Twin Networks: Matching the Future for Sequence Generation
Dmitriy Serdyuk
Nan Rosemary Ke
Alessandro Sordoni
Adam Trischler
C. Pal
Yoshua Bengio
182
12
0
22 Aug 2017
Regularizing and Optimizing LSTM Language Models
Regularizing and Optimizing LSTM Language ModelsInternational Conference on Learning Representations (ICLR), 2017
Stephen Merity
N. Keskar
R. Socher
332
1,147
0
07 Aug 2017
Revisiting Activation Regularization for Language RNNs
Revisiting Activation Regularization for Language RNNs
Stephen Merity
Bryan McCann
R. Socher
186
44
0
03 Aug 2017
Bayesian Sparsification of Recurrent Neural Networks
Bayesian Sparsification of Recurrent Neural Networks
E. Lobacheva
Nadezhda Chirkova
Dmitry Vetrov
UQCVBDL
197
16
0
31 Jul 2017
Dual Rectified Linear Units (DReLUs): A Replacement for Tanh Activation
  Functions in Quasi-Recurrent Neural Networks
Dual Rectified Linear Units (DReLUs): A Replacement for Tanh Activation Functions in Quasi-Recurrent Neural Networks
Fréderic Godin
Jonas Degrave
J. Dambre
W. D. Neve
MU
117
49
0
25 Jul 2017
Previous
1234
Next
Page 3 of 4