ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2026 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1911.00617
  4. Cited By
Explicit Explore-Exploit Algorithms in Continuous State Spaces
v1v2 (latest)

Explicit Explore-Exploit Algorithms in Continuous State Spaces

Neural Information Processing Systems (NeurIPS), 2019
1 November 2019
Mikael Henaff
    OffRL
ArXiv (abs)PDFHTML

Papers citing "Explicit Explore-Exploit Algorithms in Continuous State Spaces"

24 / 24 papers shown
Grounding Video Models to Actions through Goal Conditioned Exploration
Grounding Video Models to Actions through Goal Conditioned ExplorationInternational Conference on Learning Representations (ICLR), 2024
Yunhao Luo
Yilun Du
LM&RoVGen
401
21
0
11 Nov 2024
Exploring the Edges of Latent State Clusters for Goal-Conditioned
  Reinforcement Learning
Exploring the Edges of Latent State Clusters for Goal-Conditioned Reinforcement LearningNeural Information Processing Systems (NeurIPS), 2024
Yuanlin Duan
Guofeng Cui
He Zhu
OffRL
380
1
0
03 Nov 2024
Beyond Optimism: Exploration With Partially Observable Rewards
Beyond Optimism: Exploration With Partially Observable Rewards
Simone Parisi
Alireza Kazemipour
Michael Bowling
OffRL
216
6
0
20 Jun 2024
Planning to Go Out-of-Distribution in Offline-to-Online Reinforcement
  Learning
Planning to Go Out-of-Distribution in Offline-to-Online Reinforcement Learning
Trevor A. McInroe
Adam Jelley
Stefano V. Albrecht
Amos Storkey
OffRLOnRL
263
7
0
09 Oct 2023
Reward Model Ensembles Help Mitigate Overoptimization
Reward Model Ensembles Help Mitigate OveroptimizationInternational Conference on Learning Representations (ICLR), 2023
Thomas Coste
Usman Anwar
Robert Kirk
David M. Krueger
NoLaALM
333
181
0
04 Oct 2023
A Study of Global and Episodic Bonuses for Exploration in Contextual
  MDPs
A Study of Global and Episodic Bonuses for Exploration in Contextual MDPsInternational Conference on Machine Learning (ICML), 2023
Mikael Henaff
Minqi Jiang
Roberta Raileanu
185
15
0
05 Jun 2023
What model does MuZero learn?
What model does MuZero learn?European Conference on Artificial Intelligence (ECAI), 2023
Jinke He
Thomas M. Moerland
F. Oliehoek
333
5
0
01 Jun 2023
Conditional Mutual Information for Disentangled Representations in
  Reinforcement Learning
Conditional Mutual Information for Disentangled Representations in Reinforcement LearningNeural Information Processing Systems (NeurIPS), 2023
Mhairi Dunion
Trevor A. McInroe
K. Luck
Josiah P. Hanna
Stefano V. Albrecht
OODDRL
220
30
0
23 May 2023
Planning Goals for Exploration
Planning Goals for ExplorationInternational Conference on Learning Representations (ICLR), 2023
E. Hu
Richard Chang
Oleh Rybkin
Dinesh Jayaraman
217
30
0
23 Mar 2023
A Survey of Historical Learning: Learning Models with Learning History
A Survey of Historical Learning: Learning Models with Learning History
Xiang Li
Ge Wu
Lingfeng Yang
Wenzhe Wang
Renjie Song
Jian Yang
MUAI4TS
248
2
0
23 Mar 2023
Curiosity in Hindsight: Intrinsic Exploration in Stochastic Environments
Curiosity in Hindsight: Intrinsic Exploration in Stochastic EnvironmentsInternational Conference on Machine Learning (ICML), 2022
Daniel Jarrett
Corentin Tallec
Florent Altché
Thomas Mesnard
Rémi Munos
Michal Valko
246
6
0
18 Nov 2022
Exploration in Deep Reinforcement Learning: A Survey
Exploration in Deep Reinforcement Learning: A SurveyInformation Fusion (Inf. Fusion), 2022
Pawel Ladosz
Lilian Weng
Minwoo Kim
H. Oh
OffRL
309
493
0
02 May 2022
Generative Planning for Temporally Coordinated Exploration in
  Reinforcement Learning
Generative Planning for Temporally Coordinated Exploration in Reinforcement LearningInternational Conference on Learning Representations (ICLR), 2022
Haichao Zhang
Wei Xu
Haonan Yu
246
11
0
24 Jan 2022
Multi-Stage Episodic Control for Strategic Exploration in Text Games
Multi-Stage Episodic Control for Strategic Exploration in Text GamesInternational Conference on Learning Representations (ICLR), 2022
Jens Tuyls
Shunyu Yao
Sham Kakade
Karthik Narasimhan
296
29
0
04 Jan 2022
Explicit Explore, Exploit, or Escape ($E^4$): near-optimal
  safety-constrained reinforcement learning in polynomial time
Explicit Explore, Exploit, or Escape (E4E^4E4): near-optimal safety-constrained reinforcement learning in polynomial timeMachine-mediated learning (ML), 2021
David M. Bossens
Nick Bishop
277
8
0
14 Nov 2021
Adaptive Discretization in Online Reinforcement Learning
Adaptive Discretization in Online Reinforcement LearningOperational Research (OR), 2021
Sean R. Sinclair
Siddhartha Banerjee
Chao Yu
OffRL
252
20
0
29 Oct 2021
Imaginary Hindsight Experience Replay: Curious Model-based Learning for
  Sparse Reward Tasks
Imaginary Hindsight Experience Replay: Curious Model-based Learning for Sparse Reward Tasks
Robert McCarthy
Qiang Wang
S. Redmond
OffRL
247
16
0
05 Oct 2021
Learning to Plan Optimistically: Uncertainty-Guided Deep Exploration via
  Latent Model Ensembles
Learning to Plan Optimistically: Uncertainty-Guided Deep Exploration via Latent Model EnsemblesConference on Robot Learning (CoRL), 2020
Tim Seyde
Wilko Schwarting
S. Karaman
Daniela Rus
262
15
0
27 Oct 2020
Improved Worst-Case Regret Bounds for Randomized Least-Squares Value
  Iteration
Improved Worst-Case Regret Bounds for Randomized Least-Squares Value Iteration
Priyank Agrawal
Jinglin Chen
Nan Jiang
315
22
0
23 Oct 2020
PC-PG: Policy Cover Directed Exploration for Provable Policy Gradient
  Learning
PC-PG: Policy Cover Directed Exploration for Provable Policy Gradient LearningNeural Information Processing Systems (NeurIPS), 2020
Alekh Agarwal
Mikael Henaff
Sham Kakade
Wen Sun
OffRL
267
119
0
16 Jul 2020
Adaptive Discretization for Model-Based Reinforcement Learning
Adaptive Discretization for Model-Based Reinforcement Learning
Sean R. Sinclair
Tianyu Wang
Gauri Jain
Siddhartha Banerjee
Chao Yu
OffRL
230
24
0
01 Jul 2020
Meta-Model-Based Meta-Policy Optimization
Meta-Model-Based Meta-Policy OptimizationAsian Conference on Machine Learning (ACML), 2020
Takuya Hiraoka
Takahisa Imagawa
Voot Tangkaratt
Takayuki Osa
Takashi Onishi
Yoshimasa Tsuruoka
OffRL
407
9
0
04 Jun 2020
Planning to Explore via Self-Supervised World Models
Planning to Explore via Self-Supervised World Models
Ramanan Sekar
Oleh Rybkin
Kostas Daniilidis
Pieter Abbeel
Danijar Hafner
Deepak Pathak
SSL
335
468
0
12 May 2020
Ready Policy One: World Building Through Active Learning
Ready Policy One: World Building Through Active LearningInternational Conference on Machine Learning (ICML), 2020
Philip J. Ball
Jack Parker-Holder
Aldo Pacchiano
K. Choromanski
Stephen J. Roberts
OffRL
227
52
0
07 Feb 2020
1