ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2026 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2102.04376
  4. Cited By
Adversarially Guided Actor-Critic

Adversarially Guided Actor-Critic

International Conference on Learning Representations (ICLR), 2021
8 February 2021
Yannis Flet-Berliac
Johan Ferret
Olivier Pietquin
Philippe Preux
Matthieu Geist
ArXiv (abs)PDFHTML

Papers citing "Adversarially Guided Actor-Critic"

45 / 45 papers shown
Beyond Noisy-TVs: Noise-Robust Exploration Via Learning Progress Monitoring
Beyond Noisy-TVs: Noise-Robust Exploration Via Learning Progress Monitoring
Zhibo Hou
Zhiyu An
Wan Du
181
0
0
29 Sep 2025
CAE: Repurposing the Critic as an Explorer in Deep Reinforcement Learning
CAE: Repurposing the Critic as an Explorer in Deep Reinforcement Learning
Yexin Li
OffRL
479
2
0
23 Mar 2025
KEA: Keeping Exploration Alive by Proactively Coordinating Exploration Strategies
KEA: Keeping Exploration Alive by Proactively Coordinating Exploration Strategies
Shih-Min Yang
Martin Magnusson
J. A. Stork
Todor Stoyanov
340
1
0
23 Mar 2025
Episodic Novelty Through Temporal Distance
Episodic Novelty Through Temporal DistanceInternational Conference on Learning Representations (ICLR), 2025
Y. Jiang
Qihan Liu
Yiqin Yang
Xiaoteng Ma
Dianyu Zhong
...
Jun Yang
Bin Liang
Bo Xu
Chongjie Zhang
Qianchuan Zhao
OffRL
401
10
0
28 Jan 2025
CSAOT: Cooperative Multi-Agent System for Active Object Tracking
CSAOT: Cooperative Multi-Agent System for Active Object Tracking
Hy Nguyen
Bao Pham
Hung Du
Srikanth Thudumu
Rajesh Vasa
K. Mouzakis
230
2
0
23 Jan 2025
The impact of intrinsic rewards on exploration in Reinforcement Learning
The impact of intrinsic rewards on exploration in Reinforcement Learning
Aya Kayal
Eduardo Pignatelli
Laura Toni
305
8
0
20 Jan 2025
CALM: Curiosity-Driven Auditing for Large Language Models
CALM: Curiosity-Driven Auditing for Large Language ModelsAAAI Conference on Artificial Intelligence (AAAI), 2025
Xiang Zheng
Longxiang Wang
Yi Liu
Jie Zhang
Chao Shen
Cong Wang
MLAU
427
6
0
06 Jan 2025
Fostering Intrinsic Motivation in Reinforcement Learning with Pretrained
  Foundation Models
Fostering Intrinsic Motivation in Reinforcement Learning with Pretrained Foundation Models
Alain Andres
Javier Del Ser
OffRL
214
0
0
09 Oct 2024
NAVIX: Scaling MiniGrid Environments with JAX
NAVIX: Scaling MiniGrid Environments with JAX
Eduardo Pignatelli
Jarek Liesen
R. T. Lange
Chris Xiaoxuan Lu
Pablo Samuel Castro
Laura Toni
431
13
0
28 Jul 2024
Constrained Intrinsic Motivation for Reinforcement Learning
Constrained Intrinsic Motivation for Reinforcement Learning
Xiang Zheng
Jie Zhang
Chao Shen
Cong Wang
322
5
0
12 Jul 2024
Exploration by Learning Diverse Skills through Successor State Measures
Exploration by Learning Diverse Skills through Successor State Measures
Paul-Antoine Le Tolguenec
Yann Besse
Florent Teichteil-Königsbuch
Dennis G. Wilson
Emmanuel Rachelson
383
1
0
14 Jun 2024
Explore-Go: Leveraging Exploration for Generalisation in Deep
  Reinforcement Learning
Explore-Go: Leveraging Exploration for Generalisation in Deep Reinforcement Learning
Max Weltevrede
Felix Kaubek
M. Spaan
Wendelin Bohmer
325
0
0
12 Jun 2024
Constrained Ensemble Exploration for Unsupervised Skill Discovery
Constrained Ensemble Exploration for Unsupervised Skill Discovery
Chenjia Bai
Rushuai Yang
Qiaosheng Zhang
Kang Xu
Yi Chen
Ting Xiao
Xuelong Li
OffRL
496
9
0
25 May 2024
Joint Intrinsic Motivation for Coordinated Exploration in Multi-Agent
  Deep Reinforcement Learning
Joint Intrinsic Motivation for Coordinated Exploration in Multi-Agent Deep Reinforcement Learning
Maxime Toquebiau
Nicolas Bredèche
F. Benamar
Jae-Yun Jun
208
2
0
06 Feb 2024
Settling Decentralized Multi-Agent Coordinated Exploration by Novelty
  Sharing
Settling Decentralized Multi-Agent Coordinated Exploration by Novelty Sharing
Haobin Jiang
Haobin Jiang
Zongqing Lu
401
10
0
03 Feb 2024
Improving Intrinsic Exploration by Creating Stationary Objectives
Improving Intrinsic Exploration by Creating Stationary ObjectivesInternational Conference on Learning Representations (ICLR), 2023
Roger Creus Castanyer
Javier Civera
Taihú Pire
OffRL
496
4
0
27 Oct 2023
Intrinsic Language-Guided Exploration for Complex Long-Horizon Robotic
  Manipulation Tasks
Intrinsic Language-Guided Exploration for Complex Long-Horizon Robotic Manipulation TasksIEEE International Conference on Robotics and Automation (ICRA), 2023
Wenke Huang
Filippos Christianos
Zhibin Li
343
16
0
28 Sep 2023
Go Beyond Imagination: Maximizing Episodic Reachability with World
  Models
Go Beyond Imagination: Maximizing Episodic Reachability with World ModelsInternational Conference on Machine Learning (ICML), 2023
Yao Fu
Run Peng
Honglak Lee
234
1
0
25 Aug 2023
Diversifying AI: Towards Creative Chess with AlphaZero
Diversifying AI: Towards Creative Chess with AlphaZero
Tom Zahavy
Vivek Veeriah
Shaobo Hou
Kevin Waugh
Matthew Lai
Edouard Leurent
Nenad Tomašev
Lisa Schut
Demis Hassabis
Satinder Singh
333
23
0
17 Aug 2023
On the Importance of Exploration for Generalization in Reinforcement
  Learning
On the Importance of Exploration for Generalization in Reinforcement LearningNeural Information Processing Systems (NeurIPS), 2023
Yiding Jiang
J. Zico Kolter
Roberta Raileanu
UQCVOffRL
235
40
0
08 Jun 2023
A Study of Global and Episodic Bonuses for Exploration in Contextual
  MDPs
A Study of Global and Episodic Bonuses for Exploration in Contextual MDPsInternational Conference on Machine Learning (ICML), 2023
Mikael Henaff
Minqi Jiang
Roberta Raileanu
241
18
0
05 Jun 2023
Toward Evaluating Robustness of Reinforcement Learning with Adversarial
  Policy
Toward Evaluating Robustness of Reinforcement Learning with Adversarial PolicyDependable Systems and Networks (DSN), 2023
Jiawei Zhao
Jiabo He
Florian Schäfer
Xinyu Wang
Anima Anandkumar
Cong Wang
AAML
416
5
0
04 May 2023
Using Offline Data to Speed-up Reinforcement Learning in Procedurally
  Generated Environments
Using Offline Data to Speed-up Reinforcement Learning in Procedurally Generated EnvironmentsNeurocomputing (Neurocomputing), 2023
Alain Andres
Lukas Schafer
Esther Villar-Rodriguez
Stefano V. Albrecht
Javier Del Ser
OffRLOnRL
249
9
0
18 Apr 2023
Curiosity creates Diversity in Policy Search
Curiosity creates Diversity in Policy SearchACM Transactions on Evolutionary Learning and Optimization (TELO), 2022
Paul-Antoine Le Tolguenec
Emmanuel Rachelson
Yann Besse
Dennis G. Wilson
315
2
0
07 Dec 2022
TD3 with Reverse KL Regularizer for Offline Reinforcement Learning from
  Mixed Datasets
TD3 with Reverse KL Regularizer for Offline Reinforcement Learning from Mixed DatasetsIndustrial Conference on Data Mining (IDM), 2022
Yuanying Cai
Wei Shen
Li Zhao
Wei Shen
Xuyun Zhang
Lei Song
Jiang Bian
Tao Qin
Tie-Yan Liu
OffRL
232
6
0
05 Dec 2022
Towards Improving Exploration in Self-Imitation Learning using Intrinsic
  Motivation
Towards Improving Exploration in Self-Imitation Learning using Intrinsic MotivationIEEE Symposium Series on Computational Intelligence (IEEE SSCI), 2022
Alain Andres
Esther Villar-Rodriguez
Javier Del Ser
SSL
283
7
0
30 Nov 2022
CIM: Constrained Intrinsic Motivation for Sparse-Reward Continuous Control
Xiang Zheng
Jiabo He
Cong Wang
317
2
0
28 Nov 2022
Assessing Quality-Diversity Neuro-Evolution Algorithms Performance in
  Hard Exploration Problems
Assessing Quality-Diversity Neuro-Evolution Algorithms Performance in Hard Exploration Problems
Félix Chalumeau
Thomas Pierrot
Valentin Macé
Arthur Flajolet
Karim Beguir
Antoine Cully
Nicolas Perrin-Gilbert
467
9
0
24 Nov 2022
Curiosity in Hindsight: Intrinsic Exploration in Stochastic Environments
Curiosity in Hindsight: Intrinsic Exploration in Stochastic EnvironmentsInternational Conference on Machine Learning (ICML), 2022
Daniel Jarrett
Corentin Tallec
Florent Altché
Thomas Mesnard
Rémi Munos
Michal Valko
310
6
0
18 Nov 2022
Exploring through Random Curiosity with General Value Functions
Exploring through Random Curiosity with General Value FunctionsNeural Information Processing Systems (NeurIPS), 2022
Aditya A. Ramesh
Louis Kirsch
Sjoerd van Steenkiste
Jürgen Schmidhuber
301
13
0
18 Nov 2022
Learning General World Models in a Handful of Reward-Free Deployments
Learning General World Models in a Handful of Reward-Free DeploymentsNeural Information Processing Systems (NeurIPS), 2022
Yingchen Xu
Jack Parker-Holder
Aldo Pacchiano
Philip J. Ball
Oleh Rybkin
Stephen J. Roberts
Tim Rocktaschel
Edward Grefenstette
OffRL
301
13
0
23 Oct 2022
Exploration via Elliptical Episodic Bonuses
Exploration via Elliptical Episodic BonusesNeural Information Processing Systems (NeurIPS), 2022
Mikael Henaff
Roberta Raileanu
Minqi Jiang
Tim Rocktaschel
OffRL
407
58
0
11 Oct 2022
LECO: Learnable Episodic Count for Task-Specific Intrinsic Reward
LECO: Learnable Episodic Count for Task-Specific Intrinsic RewardNeural Information Processing Systems (NeurIPS), 2022
DaeJin Jo
Sungwoong Kim
D. W. Nam
Taehwan Kwon
Seungeun Rho
Jongmin Kim
Donghoon Lee
OffRL
181
12
0
11 Oct 2022
An Evaluation Study of Intrinsic Motivation Techniques applied to
  Reinforcement Learning over Hard Exploration Environments
An Evaluation Study of Intrinsic Motivation Techniques applied to Reinforcement Learning over Hard Exploration EnvironmentsInternational Cross-Domain Conference on Machine Learning and Knowledge Extraction (CD-MAKE), 2022
Alain Andres
Esther Villar-Rodriguez
Javier Del Ser
213
11
0
23 May 2022
$q$-Munchausen Reinforcement Learning
qqq-Munchausen Reinforcement Learning
Lingwei Zhu
Zheng Chen
E. Uchibe
Takamitsu Matsubara
OffRL
161
0
0
16 May 2022
A Hierarchical Bayesian Approach to Inverse Reinforcement Learning with
  Symbolic Reward Machines
A Hierarchical Bayesian Approach to Inverse Reinforcement Learning with Symbolic Reward MachinesInternational Conference on Machine Learning (ICML), 2022
Weichao Zhou
Wenchao Li
BDL
208
14
0
20 Apr 2022
Programmatic Reward Design by Example
Programmatic Reward Design by Example
Weichao Zhou
Wenchao Li
395
15
0
14 Dec 2021
More Efficient Exploration with Symbolic Priors on Action Sequence
  Equivalences
More Efficient Exploration with Symbolic Priors on Action Sequence Equivalences
Toby Johnstone
Nathan Grinsztajn
Johan Ferret
Philippe Preux
195
0
0
20 Oct 2021
MiniHack the Planet: A Sandbox for Open-Ended Reinforcement Learning
  Research
MiniHack the Planet: A Sandbox for Open-Ended Reinforcement Learning Research
Mikayel Samvelyan
Robert Kirk
Vitaly Kurin
Jack Parker-Holder
Minqi Jiang
Eric Hambro
Fabio Petroni
Heinrich Küttler
Edward Grefenstette
Tim Rocktaschel
OffRL
545
111
0
27 Sep 2021
Decoupled Reinforcement Learning to Stabilise Intrinsically-Motivated
  Exploration
Decoupled Reinforcement Learning to Stabilise Intrinsically-Motivated ExplorationAdaptive Agents and Multi-Agent Systems (AAMAS), 2021
Lukas Schafer
Filippos Christianos
Josiah P. Hanna
Stefano V. Albrecht
255
26
0
19 Jul 2021
Explore and Control with Adversarial Surprise
Explore and Control with Adversarial Surprise
Arnaud Fickinger
Natasha Jaques
Samyak Parajuli
Michael Chang
Nicholas Rhinehart
Glen Berseth
Stuart J. Russell
Sergey Levine
322
8
0
12 Jul 2021
MADE: Exploration via Maximizing Deviation from Explored Regions
MADE: Exploration via Maximizing Deviation from Explored RegionsNeural Information Processing Systems (NeurIPS), 2021
Tianjun Zhang
Paria Rashidinejad
Jiantao Jiao
Yuandong Tian
Joseph E. Gonzalez
Stuart J. Russell
OffRL
254
50
0
18 Jun 2021
Adversarial Intrinsic Motivation for Reinforcement Learning
Adversarial Intrinsic Motivation for Reinforcement LearningNeural Information Processing Systems (NeurIPS), 2021
Ishan Durugkar
Mauricio Tec
S. Niekum
Peter Stone
OOD
594
54
0
27 May 2021
Don't Do What Doesn't Matter: Intrinsic Motivation with Action
  Usefulness
Don't Do What Doesn't Matter: Intrinsic Motivation with Action UsefulnessInternational Joint Conference on Artificial Intelligence (IJCAI), 2021
Mathieu Seurin
Florian Strub
Philippe Preux
Olivier Pietquin
238
10
0
20 May 2021
Diversity Policy Gradient for Sample Efficient Quality-Diversity
  Optimization
Diversity Policy Gradient for Sample Efficient Quality-Diversity Optimization
Thomas Pierrot
Valentin Macé
Félix Chalumeau
Arthur Flajolet
Geoffrey Cideron
Karim Beguir
Antoine Cully
Olivier Sigaud
Nicolas Perrin-Gilbert
395
77
0
15 Jun 2020
1
Page 1 of 1