ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2010.14498
  4. Cited By
Implicit Under-Parameterization Inhibits Data-Efficient Deep
  Reinforcement Learning

Implicit Under-Parameterization Inhibits Data-Efficient Deep Reinforcement Learning

27 October 2020
Aviral Kumar
Rishabh Agarwal
Dibya Ghosh
Sergey Levine
    OffRL
ArXivPDFHTML

Papers citing "Implicit Under-Parameterization Inhibits Data-Efficient Deep Reinforcement Learning"

41 / 41 papers shown
Title
Preserving Plasticity in Continual Learning with Adaptive Linearity Injection
Preserving Plasticity in Continual Learning with Adaptive Linearity Injection
Seyed Roozbeh Razavi Rohani
Khashayar Khajavi
Wesley Chung
Mo Chen
Sharan Vaswani
CLL
AI4CE
38
0
0
14 May 2025
Learning to Sample Effective and Diverse Prompts for Text-to-Image Generation
Learning to Sample Effective and Diverse Prompts for Text-to-Image Generation
Taeyoung Yun
Dinghuai Zhang
Jinkyoo Park
Ling Pan
DiffM
86
2
0
17 Feb 2025
Audiopedia: Audio QA with Knowledge
Audiopedia: Audio QA with Knowledge
Abhirama Subramanyam Penamakuri
Kiran Chhatre
Akshat Jain
KELM
AuLLM
RALM
68
0
0
31 Dec 2024
Beyond The Rainbow: High Performance Deep Reinforcement Learning on a Desktop PC
Beyond The Rainbow: High Performance Deep Reinforcement Learning on a Desktop PC
Tyler Clark
Mark Towers
Christine Evers
Jonathon Hare
OffRL
40
0
0
06 Nov 2024
Uncovering RL Integration in SSL Loss: Objective-Specific Implications for Data-Efficient RL
Uncovering RL Integration in SSL Loss: Objective-Specific Implications for Data-Efficient RL
Ömer Veysel Çağatan
Barış Akgün
OffRL
44
0
0
22 Oct 2024
MAD-TD: Model-Augmented Data stabilizes High Update Ratio RL
MAD-TD: Model-Augmented Data stabilizes High Update Ratio RL
C. Voelcker
Marcel Hussing
Eric Eaton
Amir-massoud Farahmand
Igor Gilitschenski
46
2
0
11 Oct 2024
Neuroplastic Expansion in Deep Reinforcement Learning
Neuroplastic Expansion in Deep Reinforcement Learning
Jiashun Liu
J. Obando-Ceron
Rameswar Panda
L. Pan
47
3
0
10 Oct 2024
Don't flatten, tokenize! Unlocking the key to SoftMoE's efficacy in deep RL
Don't flatten, tokenize! Unlocking the key to SoftMoE's efficacy in deep RL
Ghada Sokar
J. Obando-Ceron
Rameswar Panda
Hugo Larochelle
Pablo Samuel Castro
MoE
171
2
0
02 Oct 2024
Scaling Offline Model-Based RL via Jointly-Optimized World-Action Model Pretraining
Scaling Offline Model-Based RL via Jointly-Optimized World-Action Model Pretraining
Jie Cheng
Ruixi Qiao
Gang Xiong
Binhua Li
Yingwei Ma
Binhua Li
Yongbin Li
Yisheng Lv
OffRL
OnRL
LM&Ro
50
3
0
01 Oct 2024
The Role of Deep Learning Regularizations on Actors in Offline RL
The Role of Deep Learning Regularizations on Actors in Offline RL
Denis Tarasov
Anja Surina
Çağlar Gülçehre
OffRL
AI4CE
73
1
0
11 Sep 2024
Frequency and Generalisation of Periodic Activation Functions in Reinforcement Learning
Frequency and Generalisation of Periodic Activation Functions in Reinforcement Learning
Augustine N. Mavor-Parker
Matthew J. Sargent
Caswell Barry
Lewis D. Griffin
Clare Lyle
52
2
0
09 Jul 2024
Is Value Functions Estimation with Classification Plug-and-play for
  Offline Reinforcement Learning?
Is Value Functions Estimation with Classification Plug-and-play for Offline Reinforcement Learning?
Denis Tarasov
Kirill Brilliantov
Dmitrii Kharlapenko
OffRL
41
2
0
10 Jun 2024
K-percent Evaluation for Lifelong RL
K-percent Evaluation for Lifelong RL
Golnaz Mesbahi
Parham Mohammad Panahi
Olya Mastikhina
Martha White
Adam White
CLL
OffRL
42
0
0
02 Apr 2024
Dissecting Deep RL with High Update Ratios: Combatting Value Divergence
Dissecting Deep RL with High Update Ratios: Combatting Value Divergence
Marcel Hussing
C. Voelcker
Igor Gilitschenski
Amir-massoud Farahmand
Eric Eaton
47
3
0
09 Mar 2024
Reset & Distill: A Recipe for Overcoming Negative Transfer in Continual
  Reinforcement Learning
Reset & Distill: A Recipe for Overcoming Negative Transfer in Continual Reinforcement Learning
Hongjoon Ahn
Jinu Hyeon
Youngmin Oh
Bosun Hwang
Taesup Moon
CLL
OnRL
39
2
0
08 Mar 2024
Overestimation, Overfitting, and Plasticity in Actor-Critic: the Bitter
  Lesson of Reinforcement Learning
Overestimation, Overfitting, and Plasticity in Actor-Critic: the Bitter Lesson of Reinforcement Learning
Michal Nauman
Michal Bortkiewicz
Piotr Milo's
Tomasz Trzciñski
M. Ostaszewski
Marek Cygan
OffRL
37
17
0
01 Mar 2024
Directions of Curvature as an Explanation for Loss of Plasticity
Directions of Curvature as an Explanation for Loss of Plasticity
Alex Lewandowski
Haruto Tanaka
Dale Schuurmans
Marlos C. Machado
32
5
0
30 Nov 2023
Maintaining Plasticity in Continual Learning via Regenerative
  Regularization
Maintaining Plasticity in Continual Learning via Regenerative Regularization
Saurabh Kumar
Henrik Marklund
Benjamin Van Roy
CLL
KELM
39
18
0
23 Aug 2023
Bigger, Better, Faster: Human-level Atari with human-level efficiency
Bigger, Better, Faster: Human-level Atari with human-level efficiency
Max Schwarzer
J. Obando-Ceron
Rameswar Panda
Marc G. Bellemare
Rishabh Agarwal
Pablo Samuel Castro
OffRL
54
85
0
30 May 2023
Understanding plasticity in neural networks
Understanding plasticity in neural networks
Clare Lyle
Zeyu Zheng
Evgenii Nikishin
Bernardo Avila-Pires
Razvan Pascanu
Will Dabney
AI4CE
45
98
0
02 Mar 2023
The Ladder in Chaos: A Simple and Effective Improvement to General DRL
  Algorithms by Policy Path Trimming and Boosting
The Ladder in Chaos: A Simple and Effective Improvement to General DRL Algorithms by Policy Path Trimming and Boosting
Hongyao Tang
Hao Fei
Jianye Hao
28
1
0
02 Mar 2023
The Dormant Neuron Phenomenon in Deep Reinforcement Learning
The Dormant Neuron Phenomenon in Deep Reinforcement Learning
Ghada Sokar
Rishabh Agarwal
Pablo Samuel Castro
Utku Evci
CLL
53
90
0
24 Feb 2023
On the Geometry of Reinforcement Learning in Continuous State and Action
  Spaces
On the Geometry of Reinforcement Learning in Continuous State and Action Spaces
Saket Tiwari
Omer Gottesman
George Konidaris
29
0
0
29 Dec 2022
Stabilizing Off-Policy Deep Reinforcement Learning from Pixels
Stabilizing Off-Policy Deep Reinforcement Learning from Pixels
Edoardo Cetin
Philip J. Ball
Steve Roberts
Oya Celiktutan
40
36
0
03 Jul 2022
Contrastive Learning as Goal-Conditioned Reinforcement Learning
Contrastive Learning as Goal-Conditioned Reinforcement Learning
Benjamin Eysenbach
Tianjun Zhang
Ruslan Salakhutdinov
Sergey Levine
SSL
OffRL
42
141
0
15 Jun 2022
Overcoming the Spectral Bias of Neural Value Approximation
Overcoming the Spectral Bias of Neural Value Approximation
Ge Yang
Anurag Ajay
Pulkit Agrawal
36
25
0
09 Jun 2022
Reincarnating Reinforcement Learning: Reusing Prior Computation to
  Accelerate Progress
Reincarnating Reinforcement Learning: Reusing Prior Computation to Accelerate Progress
Rishabh Agarwal
Max Schwarzer
Pablo Samuel Castro
Rameswar Panda
Marc G. Bellemare
OffRL
OnRL
37
63
0
03 Jun 2022
Critic Sequential Monte Carlo
Critic Sequential Monte Carlo
Vasileios Lioutas
J. Lavington
Justice Sefas
Matthew Niedoba
Yunpeng Liu
Berend Zwartsenberg
Setareh Dabiri
Frank Wood
Adam Scibior
52
7
0
30 May 2022
The Primacy Bias in Deep Reinforcement Learning
The Primacy Bias in Deep Reinforcement Learning
Evgenii Nikishin
Max Schwarzer
P. DÓro
Pierre-Luc Bacon
Rameswar Panda
OnRL
96
182
0
16 May 2022
Understanding and Preventing Capacity Loss in Reinforcement Learning
Understanding and Preventing Capacity Loss in Reinforcement Learning
Clare Lyle
Mark Rowland
Will Dabney
CLL
41
110
0
20 Apr 2022
When Should We Prefer Offline Reinforcement Learning Over Behavioral
  Cloning?
When Should We Prefer Offline Reinforcement Learning Over Behavioral Cloning?
Aviral Kumar
Joey Hong
Anika Singh
Sergey Levine
OffRL
50
77
0
12 Apr 2022
Automated Reinforcement Learning (AutoRL): A Survey and Open Problems
Automated Reinforcement Learning (AutoRL): A Survey and Open Problems
Jack Parker-Holder
Raghunandan Rajan
Xingyou Song
André Biedenkapp
Yingjie Miao
...
Vu-Linh Nguyen
Roberto Calandra
Aleksandra Faust
Frank Hutter
Marius Lindauer
AI4CE
35
100
0
11 Jan 2022
DR3: Value-Based Deep Reinforcement Learning Requires Explicit
  Regularization
DR3: Value-Based Deep Reinforcement Learning Requires Explicit Regularization
Aviral Kumar
Rishabh Agarwal
Tengyu Ma
Aaron Courville
George Tucker
Sergey Levine
OffRL
31
65
0
09 Dec 2021
The Difficulty of Passive Learning in Deep Reinforcement Learning
The Difficulty of Passive Learning in Deep Reinforcement Learning
Georg Ostrovski
Pablo Samuel Castro
Will Dabney
OffRL
24
57
0
26 Oct 2021
A Workflow for Offline Model-Free Robotic Reinforcement Learning
A Workflow for Offline Model-Free Robotic Reinforcement Learning
Aviral Kumar
Anika Singh
Stephen Tian
Chelsea Finn
Sergey Levine
OffRL
143
85
0
22 Sep 2021
Modularity in Reinforcement Learning via Algorithmic Independence in
  Credit Assignment
Modularity in Reinforcement Learning via Algorithmic Independence in Credit Assignment
Michael Chang
Sid Kaushik
Sergey Levine
Thomas Griffiths
31
8
0
28 Jun 2021
Flow Network based Generative Models for Non-Iterative Diverse Candidate
  Generation
Flow Network based Generative Models for Non-Iterative Diverse Candidate Generation
Emmanuel Bengio
Moksh Jain
Maksym Korablyov
Doina Precup
Yoshua Bengio
49
312
0
08 Jun 2021
Spectral Normalisation for Deep Reinforcement Learning: an Optimisation
  Perspective
Spectral Normalisation for Deep Reinforcement Learning: an Optimisation Perspective
Florin Gogianu
Tudor Berariu
Mihaela Rosca
Claudia Clopath
L. Buşoniu
Razvan Pascanu
24
54
0
11 May 2021
Training Larger Networks for Deep Reinforcement Learning
Training Larger Networks for Deep Reinforcement Learning
Keita Ota
Devesh K. Jha
Asako Kanezaki
OffRL
37
39
0
16 Feb 2021
Learning Off-Policy with Online Planning
Learning Off-Policy with Online Planning
Harshit S. Sikchi
Wenxuan Zhou
David Held
OffRL
37
46
0
23 Aug 2020
Offline Reinforcement Learning: Tutorial, Review, and Perspectives on
  Open Problems
Offline Reinforcement Learning: Tutorial, Review, and Perspectives on Open Problems
Sergey Levine
Aviral Kumar
George Tucker
Justin Fu
OffRL
GP
343
1,968
0
04 May 2020
1