ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2026 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1908.02388
  4. Cited By
Benchmarking Bonus-Based Exploration Methods on the Arcade Learning
  Environment
v1v2v3 (latest)

Benchmarking Bonus-Based Exploration Methods on the Arcade Learning Environment

6 August 2019
Adrien Ali Taïga
W. Fedus
Marlos C. Machado
Aaron Courville
Marc G. Bellemare
ArXiv (abs)PDFHTML

Papers citing "Benchmarking Bonus-Based Exploration Methods on the Arcade Learning Environment"

28 / 28 papers shown
Sim-to-Real Reinforcement Learning for Vision-Based Dexterous Manipulation on Humanoids
Sim-to-Real Reinforcement Learning for Vision-Based Dexterous Manipulation on Humanoids
Toru Lin
Kartik Sachdev
Linxi Fan
Jitendra Malik
Yuke Zhu
521
64
0
27 Feb 2025
The Cross-environment Hyperparameter Setting Benchmark for Reinforcement
  Learning
The Cross-environment Hyperparameter Setting Benchmark for Reinforcement Learning
Andrew Patterson
Samuel Neumann
Raksha Kumaraswamy
Martha White
Adam White
292
3
0
26 Jul 2024
Random Latent Exploration for Deep Reinforcement Learning
Random Latent Exploration for Deep Reinforcement Learning
Srinath Mahankali
Zhang-Wei Hong
Ayush Sekhari
Alexander Rakhlin
Pulkit Agrawal
723
8
0
18 Jul 2024
Offline Retraining for Online RL: Decoupled Policy Learning to Mitigate
  Exploration Bias
Offline Retraining for Online RL: Decoupled Policy Learning to Mitigate Exploration Bias
Max Sobol Mark
Archit Sharma
Fahim Tajwar
Rafael Rafailov
Sergey Levine
Chelsea Finn
OffRLOnRL
342
4
0
12 Oct 2023
Towards practical reinforcement learning for tokamak magnetic control
Towards practical reinforcement learning for tokamak magnetic controlFusion engineering and design (FED), 2023
Brendan D. Tracey
Andrea Michi
Yuri Chervonyi
Ian Davies
Cosmin Paduraru
...
Jonathan Evens
Paula Kurylowicz
D. Mankowitz
Martin Riedmiller
The Tcv Team
AI4CE
250
20
0
21 Jul 2023
Deep Reinforcement Learning with Plasticity Injection
Deep Reinforcement Learning with Plasticity InjectionNeural Information Processing Systems (NeurIPS), 2023
Evgenii Nikishin
Junhyuk Oh
Georg Ostrovski
Clare Lyle
Razvan Pascanu
Will Dabney
André Barreto
OffRL
380
72
0
24 May 2023
MIMEx: Intrinsic Rewards from Masked Input Modeling
MIMEx: Intrinsic Rewards from Masked Input ModelingNeural Information Processing Systems (NeurIPS), 2023
Toru Lin
Allan Jabri
OffRL
307
7
0
15 May 2023
Empirical Design in Reinforcement Learning
Empirical Design in Reinforcement Learning
Andrew Patterson
Samuel Neumann
Martha White
Adam White
389
52
0
03 Apr 2023
Conditionally Optimistic Exploration for Cooperative Deep Multi-Agent
  Reinforcement Learning
Conditionally Optimistic Exploration for Cooperative Deep Multi-Agent Reinforcement LearningConference on Uncertainty in Artificial Intelligence (UAI), 2023
Xutong Zhao
Yangchen Pan
Chenjun Xiao
Sarath Chandar
Janarthanan Rajendran
332
9
0
16 Mar 2023
Redeeming Intrinsic Rewards via Constrained Optimization
Redeeming Intrinsic Rewards via Constrained OptimizationNeural Information Processing Systems (NeurIPS), 2022
Eric Chen
Zhang-Wei Hong
Joni Pajarinen
Pulkit Agrawal
OnRL
317
38
0
14 Nov 2022
Collaborative Training of Heterogeneous Reinforcement Learning Agents in
  Environments with Sparse Rewards: What and When to Share?
Collaborative Training of Heterogeneous Reinforcement Learning Agents in Environments with Sparse Rewards: What and When to Share?
Alain Andres
Esther Villar-Rodriguez
Javier Del Ser
330
12
0
24 Feb 2022
Hit and Lead Discovery with Explorative RL and Fragment-based Molecule
  Generation
Hit and Lead Discovery with Explorative RL and Fragment-based Molecule Generation
Soojung Yang
Doyeong Hwang
Seul Lee
Seongok Ryu
Sung Ju Hwang
520
94
0
04 Oct 2021
Decoupled Reinforcement Learning to Stabilise Intrinsically-Motivated
  Exploration
Decoupled Reinforcement Learning to Stabilise Intrinsically-Motivated ExplorationAdaptive Agents and Multi-Agent Systems (AAMAS), 2021
Lukas Schafer
Filippos Christianos
Josiah P. Hanna
Stefano V. Albrecht
255
26
0
19 Jul 2021
Tuning Mixed Input Hyperparameters on the Fly for Efficient Population
  Based AutoRL
Tuning Mixed Input Hyperparameters on the Fly for Efficient Population Based AutoRLNeural Information Processing Systems (NeurIPS), 2021
Jack Parker-Holder
Vu Nguyen
Shaan Desai
Stephen J. Roberts
235
20
0
30 Jun 2021
Policy Finetuning: Bridging Sample-Efficient Offline and Online
  Reinforcement Learning
Policy Finetuning: Bridging Sample-Efficient Offline and Online Reinforcement LearningNeural Information Processing Systems (NeurIPS), 2021
Tengyang Xie
Nan Jiang
Huan Wang
Caiming Xiong
Yu Bai
OffRLOnRL
392
188
0
09 Jun 2021
Learning on a Budget via Teacher Imitation
Learning on a Budget via Teacher Imitation
Ercüment Ilhan
Jeremy Gow
Diego Perez-Liebana
OffRL
229
2
0
17 Apr 2021
State Entropy Maximization with Random Encoders for Efficient
  Exploration
State Entropy Maximization with Random Encoders for Efficient ExplorationInternational Conference on Machine Learning (ICML), 2021
Younggyo Seo
Lili Chen
Jinwoo Shin
Honglak Lee
Pieter Abbeel
Kimin Lee
349
155
0
18 Feb 2021
Locally Persistent Exploration in Continuous Control Tasks with Sparse
  Rewards
Locally Persistent Exploration in Continuous Control Tasks with Sparse RewardsInternational Conference on Machine Learning (ICML), 2020
Susan Amin
Maziar Gomrokchi
Hossein Aboutalebi
Harsh Satija
Doina Precup
239
17
0
26 Dec 2020
Self-Imitation Advantage Learning
Self-Imitation Advantage LearningAdaptive Agents and Multi-Agent Systems (AAMAS), 2020
Johan Ferret
Olivier Pietquin
Matthieu Geist
291
24
0
22 Dec 2020
Student-Initiated Action Advising via Advice Novelty
Student-Initiated Action Advising via Advice NoveltyIEEE Transactions on Games (IEEE Trans. Games), 2020
Ercüment Ilhan
Jeremy Gow
Diego Perez
151
10
0
01 Oct 2020
See, Hear, Explore: Curiosity via Audio-Visual Association
See, Hear, Explore: Curiosity via Audio-Visual AssociationNeural Information Processing Systems (NeurIPS), 2020
Victoria Dean
Shubham Tulsiani
Abhinav Gupta
329
64
0
07 Jul 2020
Meta-learning curiosity algorithms
Meta-learning curiosity algorithmsInternational Conference on Learning Representations (ICLR), 2020
Ferran Alet
Martin Schneider
Tomas Lozano-Perez
L. Kaelbling
284
67
0
11 Mar 2020
Exploring Unknown States with Action Balance
Exploring Unknown States with Action Balance
Yan Song
Yingfeng Chen
Yujing Hu
Changjie Fan
205
6
0
10 Mar 2020
On Catastrophic Interference in Atari 2600 Games
On Catastrophic Interference in Atari 2600 Games
W. Fedus
Dibya Ghosh
John D. Martin
Marc G. Bellemare
Yoshua Bengio
Hugo Larochelle
201
28
0
28 Feb 2020
Making Efficient Use of Demonstrations to Solve Hard Exploration
  Problems
Making Efficient Use of Demonstrations to Solve Hard Exploration ProblemsInternational Conference on Learning Representations (ICLR), 2019
T. Paine
Çağlar Gülçehre
Bobak Shahriari
Misha Denil
Matt Hoffman
...
Duncan Williams
Gabriel Barth-Maron
Ziyun Wang
Nando de Freitas
Worlds Team
198
96
0
03 Sep 2019
A survey on intrinsic motivation in reinforcement learning
A survey on intrinsic motivation in reinforcement learning
A. Aubret
L. Matignon
S. Hassas
AI4CE
593
163
0
19 Aug 2019
Efficient Exploration via State Marginal Matching
Efficient Exploration via State Marginal Matching
Lisa Lee
Benjamin Eysenbach
Emilio Parisotto
Eric Xing
Sergey Levine
Ruslan Salakhutdinov
548
282
0
12 Jun 2019
Count-Based Exploration with the Successor Representation
Count-Based Exploration with the Successor Representation
Marlos C. Machado
Marc G. Bellemare
Michael Bowling
323
205
0
31 Jul 2018
1
Page 1 of 1