ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1604.01946
  4. Cited By
Optimizing Performance of Recurrent Neural Networks on GPUs

Optimizing Performance of Recurrent Neural Networks on GPUs

7 April 2016
J. Appleyard
Tomás Kociský
Phil Blunsom
ArXivPDFHTML

Papers citing "Optimizing Performance of Recurrent Neural Networks on GPUs"

12 / 12 papers shown
Title
Simple Recurrence Improves Masked Language Models
Simple Recurrence Improves Masked Language Models
Tao Lei
Ran Tian
Jasmijn Bastings
Ankur P. Parikh
77
4
0
23 May 2022
Privacy-preserving Federated Learning for Residential Short Term Load
  Forecasting
Privacy-preserving Federated Learning for Residential Short Term Load Forecasting
Joaquín Delgado Fernández
Sergio Potenciano Menci
Chul Min Lee
Gilbert Fridgen
28
53
0
17 Nov 2021
A Distributed Deep Reinforcement Learning Technique for Application
  Placement in Edge and Fog Computing Environments
A Distributed Deep Reinforcement Learning Technique for Application Placement in Edge and Fog Computing Environments
M. Goudarzi
M. Palaniswami
Rajkumar Buyya
OffRL
27
85
0
24 Oct 2021
SMILES-X: autonomous molecular compounds characterization for small
  datasets without descriptors
SMILES-X: autonomous molecular compounds characterization for small datasets without descriptors
G. Lambard
Ekaterina Gracheva
19
20
0
20 Jun 2019
A Lightweight Recurrent Network for Sequence Modeling
A Lightweight Recurrent Network for Sequence Modeling
Biao Zhang
Rico Sennrich
27
7
0
30 May 2019
Scheduling Computation Graphs of Deep Learning Models on Manycore CPUs
Scheduling Computation Graphs of Deep Learning Models on Manycore CPUs
Linpeng Tang
Yida Wang
Theodore L. Willke
Kai Li
GNN
11
22
0
16 Jul 2018
LSTM Benchmarks for Deep Learning Frameworks
LSTM Benchmarks for Deep Learning Frameworks
Stefan Braun
20
28
0
05 Jun 2018
Echo: Compiler-based GPU Memory Footprint Reduction for LSTM RNN
  Training
Echo: Compiler-based GPU Memory Footprint Reduction for LSTM RNN Training
Bojian Zheng
Abhishek Tiwari
Nandita Vijaykumar
Gennady Pekhimenko
19
44
0
22 May 2018
Demystifying Parallel and Distributed Deep Learning: An In-Depth
  Concurrency Analysis
Demystifying Parallel and Distributed Deep Learning: An In-Depth Concurrency Analysis
Tal Ben-Nun
Torsten Hoefler
GNN
24
701
0
26 Feb 2018
A Scalable Near-Memory Architecture for Training Deep Neural Networks on
  Large In-Memory Datasets
A Scalable Near-Memory Architecture for Training Deep Neural Networks on Large In-Memory Datasets
Fabian Schuiki
Michael Schaffner
Frank K. Gürkaynak
Luca Benini
21
70
0
19 Feb 2018
IMPALA: Scalable Distributed Deep-RL with Importance Weighted
  Actor-Learner Architectures
IMPALA: Scalable Distributed Deep-RL with Importance Weighted Actor-Learner Architectures
L. Espeholt
Hubert Soyer
Rémi Munos
Karen Simonyan
Volodymyr Mnih
...
Vlad Firoiu
Tim Harley
Iain Dunning
Shane Legg
Koray Kavukcuoglu
13
1,569
0
05 Feb 2018
MobiRNN: Efficient Recurrent Neural Network Execution on Mobile GPU
MobiRNN: Efficient Recurrent Neural Network Execution on Mobile GPU
Qingqing Cao
Niranjan Balasubramanian
A. Balasubramanian
18
61
0
03 Jun 2017
1