ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2026 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1203.1007
  4. Cited By
Agnostic System Identification for Model-Based Reinforcement Learning
v1v2 (latest)

Agnostic System Identification for Model-Based Reinforcement Learning

International Conference on Machine Learning (ICML), 2012
5 March 2012
Stéphane Ross
Drew Bagnell
ArXiv (abs)PDFHTML

Papers citing "Agnostic System Identification for Model-Based Reinforcement Learning"

50 / 76 papers shown
Multi-agent Coordination via Flow Matching
Multi-agent Coordination via Flow Matching
Dongsu Lee
Daehee Lee
Amy Zhang
191
1
0
07 Nov 2025
Finite-Time Bounds for Average-Reward Fitted Q-Iteration
Finite-Time Bounds for Average-Reward Fitted Q-Iteration
Jongmin Lee
Ernest K. Ryu
OffRL
135
0
0
20 Oct 2025
Offline vs. Online Learning in Model-based RL: Lessons for Data Collection Strategies
Offline vs. Online Learning in Model-based RL: Lessons for Data Collection Strategies
Jiaqi Chen
Ji Shi
Cansu Sancaktar
Jonas Frey
Georg Martius
OffRL
153
0
0
06 Sep 2025
Augmenting Online RL with Offline Data is All You Need: A Unified Hybrid RL Algorithm Design and Analysis
Augmenting Online RL with Offline Data is All You Need: A Unified Hybrid RL Algorithm Design and Analysis
Ruiquan Huang
Donghao Li
Chengshuai Shi
Cong Shen
Jing Yang
OffRL
503
0
0
01 Jul 2025
A Smooth Sea Never Made a Skilled SAILOR: Robust Imitation via Learning to Search
A Smooth Sea Never Made a Skilled SAILOR: Robust Imitation via Learning to Search
A. Jain
Vibhakar Mohta
Subin Kim
Atiksh Bhardwaj
Juntao Ren
Yunhai Feng
Sanjiban Choudhury
Gokul Swamy
OffRL
437
14
0
05 Jun 2025
Trajectory First: A Curriculum for Discovering Diverse Policies
Trajectory First: A Curriculum for Discovering Diverse Policies
Cornelius V. Braun
Sayantan Auddy
Marc Toussaint
372
1
0
02 Jun 2025
Policy-Driven World Model Adaptation for Robust Offline Model-based Reinforcement Learning
Policy-Driven World Model Adaptation for Robust Offline Model-based Reinforcement Learning
Jiayu Chen
Aravind Venugopal
Shiyu Huang
Jeff Schneider
OffRL
521
0
0
19 May 2025
ACL-QL: Adaptive Conservative Level in Q-Learning for Offline Reinforcement Learning
ACL-QL: Adaptive Conservative Level in Q-Learning for Offline Reinforcement LearningIEEE Transactions on Neural Networks and Learning Systems (TNNLS), 2024
Kun Wu
Yinuo Zhao
Zhihao Xu
Zhengping Che
Chengxiang Yin
C. Liu
Qinru Qiu
Feiferi Feng
OffRL
501
10
0
22 Dec 2024
Hybrid Reinforcement Learning from Offline Observation Alone
Hybrid Reinforcement Learning from Offline Observation Alone
Yuda Song
J. Andrew Bagnell
Aarti Singh
OffRL
348
6
0
11 Jun 2024
The Virtues of Pessimism in Inverse Reinforcement Learning
David Wu
Gokul Swamy
J. Andrew Bagnell
Zhiwei Steven Wu
Sanjiban Choudhury
369
0
0
04 Feb 2024
MoMA: Model-based Mirror Ascent for Offline Reinforcement Learning
MoMA: Model-based Mirror Ascent for Offline Reinforcement Learning
Mao Hong
Zhiyue Zhang
Yue Wu
Yan Xu
OffRL
303
1
0
21 Jan 2024
Boosting Reinforcement Learning and Planning with Demonstrations: A
  Survey
Boosting Reinforcement Learning and Planning with Demonstrations: A Survey
Tongzhou Mu
H. Su
OffRL
454
1
0
23 Mar 2023
On the Sample Complexity of Vanilla Model-Based Offline Reinforcement
  Learning with Dependent Samples
On the Sample Complexity of Vanilla Model-Based Offline Reinforcement Learning with Dependent SamplesAAAI Conference on Artificial Intelligence (AAAI), 2023
Mustafa O. Karabag
Ufuk Topcu
OffRL
339
6
0
07 Mar 2023
The Virtues of Laziness in Model-based RL: A Unified Objective and
  Algorithms
The Virtues of Laziness in Model-based RL: A Unified Objective and AlgorithmsInternational Conference on Machine Learning (ICML), 2023
Anirudh Vemula
Yuda Song
Aarti Singh
J. Andrew Bagnell
Sanjiban Choudhury
OffRL
292
15
0
01 Mar 2023
Predictable MDP Abstraction for Unsupervised Model-Based RL
Predictable MDP Abstraction for Unsupervised Model-Based RLInternational Conference on Machine Learning (ICML), 2023
Seohong Park
Sergey Levine
312
11
0
08 Feb 2023
Efficient Online Reinforcement Learning with Offline Data
Efficient Online Reinforcement Learning with Offline DataInternational Conference on Machine Learning (ICML), 2023
Philip J. Ball
Laura M. Smith
Ilya Kostrikov
Sergey Levine
OffRLOnRL
697
310
0
06 Feb 2023
Policy Expansion for Bridging Offline-to-Online Reinforcement Learning
Policy Expansion for Bridging Offline-to-Online Reinforcement LearningInternational Conference on Learning Representations (ICLR), 2023
Haichao Zhang
Weiwen Xu
Haonan Yu
OffRLOnRLCLL
630
104
0
02 Feb 2023
Selective Uncertainty Propagation in Offline RL
Selective Uncertainty Propagation in Offline RLAAAI Conference on Artificial Intelligence (AAAI), 2023
Sanath Kumar Krishnamurthy
Shrey Modi
Tanmay Gangwani
S. Katariya
Branislav Kveton
A. Rangi
OffRL
727
0
0
01 Feb 2023
Leveraging Offline Data in Online Reinforcement Learning
Leveraging Offline Data in Online Reinforcement LearningInternational Conference on Machine Learning (ICML), 2022
Andrew Wagenmaker
Aldo Pacchiano
OffRLOnRL
411
46
0
09 Nov 2022
Hybrid RL: Using Both Offline and Online Data Can Make RL Efficient
Hybrid RL: Using Both Offline and Online Data Can Make RL EfficientInternational Conference on Learning Representations (ICLR), 2022
Yuda Song
Yi Zhou
Ayush Sekhari
J. Andrew Bagnell
A. Krishnamurthy
Wen Sun
OffRLOnRL
458
142
0
13 Oct 2022
A Unified Framework for Alternating Offline Model Training and Policy
  Learning
A Unified Framework for Alternating Offline Model Training and Policy LearningNeural Information Processing Systems (NeurIPS), 2022
Shentao Yang
Shujian Zhang
Yihao Feng
Mi Zhou
OffRL
322
17
0
12 Oct 2022
Relational Reasoning via Set Transformers: Provable Efficiency and
  Applications to MARL
Relational Reasoning via Set Transformers: Provable Efficiency and Applications to MARLNeural Information Processing Systems (NeurIPS), 2022
Fengzhuo Zhang
Boyi Liu
Kaixin Wang
Vincent Y. F. Tan
Zhuoran Yang
Zhaoran Wang
OffRLLRM
326
14
0
20 Sep 2022
Conservative Dual Policy Optimization for Efficient Model-Based
  Reinforcement Learning
Conservative Dual Policy Optimization for Efficient Model-Based Reinforcement LearningNeural Information Processing Systems (NeurIPS), 2022
Shen Zhang
215
6
0
16 Sep 2022
Double Check Your State Before Trusting It: Confidence-Aware
  Bidirectional Offline Model-Based Imagination
Double Check Your State Before Trusting It: Confidence-Aware Bidirectional Offline Model-Based ImaginationNeural Information Processing Systems (NeurIPS), 2022
Jiafei Lyu
Xiu Li
Zongqing Lu
OffRL
389
32
0
16 Jun 2022
Regularizing a Model-based Policy Stationary Distribution to Stabilize
  Offline Reinforcement Learning
Regularizing a Model-based Policy Stationary Distribution to Stabilize Offline Reinforcement LearningInternational Conference on Machine Learning (ICML), 2022
Shentao Yang
Yihao Feng
Shujian Zhang
Mi Zhou
OffRL
289
14
0
14 Jun 2022
Online No-regret Model-Based Meta RL for Personalized Navigation
Online No-regret Model-Based Meta RL for Personalized NavigationConference on Learning for Dynamics & Control (L4DC), 2022
Yuda Song
Ye Yuan
Wen Sun
Kris Kitani
265
0
0
05 Apr 2022
Value Gradient weighted Model-Based Reinforcement Learning
Value Gradient weighted Model-Based Reinforcement LearningInternational Conference on Learning Representations (ICLR), 2022
C. Voelcker
Victor Liao
Animesh Garg
Amir-massoud Farahmand
250
36
0
04 Apr 2022
AKF-SR: Adaptive Kalman Filtering-based Successor Representation
AKF-SR: Adaptive Kalman Filtering-based Successor Representation
Parvin Malekzadeh
Mohammad Salimibeni
Ming Hou
Arash Mohammadi
Konstantinos N. Plataniotis
297
6
0
31 Mar 2022
How to Leverage Unlabeled Data in Offline Reinforcement Learning
How to Leverage Unlabeled Data in Offline Reinforcement LearningInternational Conference on Machine Learning (ICML), 2022
Tianhe Yu
Aviral Kumar
Yevgen Chebotar
Karol Hausman
Chelsea Finn
Sergey Levine
OffRL
605
78
0
03 Feb 2022
Offline Reinforcement Learning: Fundamental Barriers for Value Function
  Approximation
Offline Reinforcement Learning: Fundamental Barriers for Value Function ApproximationAnnual Conference Computational Learning Theory (COLT), 2021
Dylan J. Foster
A. Krishnamurthy
D. Simchi-Levi
Yunzong Xu
OffRL
358
74
0
21 Nov 2021
Mismatched No More: Joint Model-Policy Optimization for Model-Based RL
Mismatched No More: Joint Model-Policy Optimization for Model-Based RL
Benjamin Eysenbach
Alexander Khazatsky
Sergey Levine
Ruslan Salakhutdinov
OffRL
556
52
0
06 Oct 2021
DROMO: Distributionally Robust Offline Model-based Policy Optimization
DROMO: Distributionally Robust Offline Model-based Policy Optimization
Ruizhen Liu
Dazhi Zhong
Zhi-Cong Chen
OffRL
223
3
0
15 Sep 2021
Non-Markovian Reinforcement Learning using Fractional Dynamics
Non-Markovian Reinforcement Learning using Fractional DynamicsIEEE Conference on Decision and Control (CDC), 2021
Gaurav Gupta
Chenzhong Yin
Jyotirmoy V. Deshmukh
P. Bogdan
OffRL
204
16
0
29 Jul 2021
PC-MLP: Model-based Reinforcement Learning with Policy Cover Guided
  Exploration
PC-MLP: Model-based Reinforcement Learning with Policy Cover Guided ExplorationInternational Conference on Machine Learning (ICML), 2021
Yuda Song
Wen Sun
309
23
0
15 Jul 2021
Pessimistic Model-based Offline Reinforcement Learning under Partial
  Coverage
Pessimistic Model-based Offline Reinforcement Learning under Partial Coverage
Masatoshi Uehara
Wen Sun
OffRL
528
169
0
13 Jul 2021
Identity Concealment Games: How I Learned to Stop Revealing and Love the
  Coincidences
Identity Concealment Games: How I Learned to Stop Revealing and Love the Coincidences
Mustafa O. Karabag
Melkior Ornik
Ufuk Topcu
293
4
0
12 May 2021
Instabilities of Offline RL with Pre-Trained Neural Representation
Instabilities of Offline RL with Pre-Trained Neural RepresentationInternational Conference on Machine Learning (ICML), 2021
Ruosong Wang
Yifan Wu
Ruslan Salakhutdinov
Sham Kakade
OffRL
393
45
0
08 Mar 2021
COMBO: Conservative Offline Model-Based Policy Optimization
COMBO: Conservative Offline Model-Based Policy OptimizationNeural Information Processing Systems (NeurIPS), 2021
Tianhe Yu
Aviral Kumar
Rafael Rafailov
Aravind Rajeswaran
Sergey Levine
Chelsea Finn
OffRL
768
505
0
16 Feb 2021
Blending MPC & Value Function Approximation for Efficient Reinforcement
  Learning
Blending MPC & Value Function Approximation for Efficient Reinforcement LearningInternational Conference on Learning Representations (ICLR), 2020
M. Bhardwaj
Sanjiban Choudhury
Byron Boots
370
45
0
10 Dec 2020
What are the Statistical Limits of Offline RL with Linear Function
  Approximation?
What are the Statistical Limits of Offline RL with Linear Function Approximation?
Ruosong Wang
Dean Phillips Foster
Sham Kakade
OffRL
464
171
0
22 Oct 2020
Driving Through Ghosts: Behavioral Cloning with False Positives
Driving Through Ghosts: Behavioral Cloning with False PositivesIEEE/RJS International Conference on Intelligent RObots and Systems (IROS), 2020
Andreas Buhler
Adrien Gaidon
Andrei Cramariuc
Rares Andrei Ambrus
Guy Rosman
Wolfram Burgard
265
18
0
29 Aug 2020
Explaining Fast Improvement in Online Imitation Learning
Explaining Fast Improvement in Online Imitation Learning
Xinyan Yan
Byron Boots
Ching-An Cheng
OnRL
251
1
0
06 Jul 2020
Off-Dynamics Reinforcement Learning: Training for Transfer with Domain
  Classifiers
Off-Dynamics Reinforcement Learning: Training for Transfer with Domain ClassifiersInternational Conference on Learning Representations (ICLR), 2020
Benjamin Eysenbach
Swapnil Asawa
Shreyas Chaudhari
Sergey Levine
Ruslan Salakhutdinov
462
120
0
24 Jun 2020
Neural Dynamical Systems: Balancing Structure and Flexibility in
  Physical Prediction
Neural Dynamical Systems: Balancing Structure and Flexibility in Physical PredictionIEEE Conference on Decision and Control (CDC), 2020
Viraj Mehta
I. Char
Willie Neiswanger
Youngseog Chung
A. Nelson
M. Boyer
E. Kolemen
J. Schneider
AI4CE
218
33
0
23 Jun 2020
Information Theoretic Regret Bounds for Online Nonlinear Control
Information Theoretic Regret Bounds for Online Nonlinear Control
Sham Kakade
A. Krishnamurthy
Kendall Lowrey
Motoya Ohnishi
Wen Sun
303
132
0
22 Jun 2020
Active Learning for Nonlinear System Identification with Guarantees
Active Learning for Nonlinear System Identification with Guarantees
Horia Mania
Sai Li
Benjamin Recht
269
123
0
18 Jun 2020
Provably Efficient Model-based Policy Adaptation
Provably Efficient Model-based Policy AdaptationInternational Conference on Machine Learning (ICML), 2020
Yuda Song
Aditi Mavalankar
Wen Sun
Sicun Gao
TTAOffRL
239
11
0
14 Jun 2020
Learning Active Task-Oriented Exploration Policies for Bridging the
  Sim-to-Real Gap
Learning Active Task-Oriented Exploration Policies for Bridging the Sim-to-Real Gap
Jacky Liang
Saumya Saxena
Oliver Kroemer
307
23
0
02 Jun 2020
MM-KTD: Multiple Model Kalman Temporal Differences for Reinforcement
  Learning
MM-KTD: Multiple Model Kalman Temporal Differences for Reinforcement LearningIEEE Access (IEEE Access), 2020
Parvin Malekzadeh
Mohammad Salimibeni
Arash Mohammadi
A. Assa
Konstantinos N. Plataniotis
OffRL
132
13
0
30 May 2020
MOReL : Model-Based Offline Reinforcement Learning
MOReL : Model-Based Offline Reinforcement Learning
Rahul Kidambi
Aravind Rajeswaran
Praneeth Netrapalli
Thorsten Joachims
OffRL
555
793
0
12 May 2020
12
Next
Page 1 of 2