ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2205.10629
  4. Cited By
User-Interactive Offline Reinforcement Learning

User-Interactive Offline Reinforcement Learning

21 May 2022
Phillip Swazinna
Steffen Udluft
Thomas Runkler
    OffRL
ArXivPDFHTML

Papers citing "User-Interactive Offline Reinforcement Learning"

10 / 10 papers shown
Title
TEA: Trajectory Encoding Augmentation for Robust and Transferable Policies in Offline Reinforcement Learning
TEA: Trajectory Encoding Augmentation for Robust and Transferable Policies in Offline Reinforcement Learning
Batıkan Bora Ormancı
Phillip Swazinna
Steffen Udluft
Thomas Runkler
OffRL
66
0
0
28 Jan 2025
Preference-Optimized Pareto Set Learning for Blackbox Optimization
Preference-Optimized Pareto Set Learning for Blackbox Optimization
Zhang Haishan
Diptesh Das
Koji Tsuda
24
1
0
19 Aug 2024
Debiased Offline Representation Learning for Fast Online Adaptation in Non-stationary Dynamics
Debiased Offline Representation Learning for Fast Online Adaptation in Non-stationary Dynamics
Xinyu Zhang
Wenjie Qiu
Yi-Chen Li
Lei Yuan
Chengxing Jia
Zongzhang Zhang
Yang Yu
OffRL
20
1
0
17 Feb 2024
Train Once, Get a Family: State-Adaptive Balances for Offline-to-Online
  Reinforcement Learning
Train Once, Get a Family: State-Adaptive Balances for Offline-to-Online Reinforcement Learning
Shenzhi Wang
Qisen Yang
Jiawei Gao
Matthieu Lin
Hao Chen
Liwei Wu
Ning Jia
Shiji Song
Gao Huang
OffRL
16
12
0
27 Oct 2023
Learning Control Policies for Variable Objectives from Offline Data
Learning Control Policies for Variable Objectives from Offline Data
Marc Weber
Phillip Swazinna
D. Hein
Steffen Udluft
V. Sterzing
OffRL
8
8
0
11 Aug 2023
Automatic Trade-off Adaptation in Offline RL
Automatic Trade-off Adaptation in Offline RL
Phillip Swazinna
Steffen Udluft
Thomas Runkler
OffRL
11
0
0
16 Jun 2023
Value function estimation using conditional diffusion models for control
Value function estimation using conditional diffusion models for control
Bogdan Mazoure
Walter A. Talbott
Miguel Angel Bautista
R. Devon Hjelm
Alexander Toshev
J. Susskind
DiffM
12
4
0
09 Jun 2023
Offline Reinforcement Learning with Implicit Q-Learning
Offline Reinforcement Learning with Implicit Q-Learning
Ilya Kostrikov
Ashvin Nair
Sergey Levine
OffRL
212
832
0
12 Oct 2021
COMBO: Conservative Offline Model-Based Policy Optimization
COMBO: Conservative Offline Model-Based Policy Optimization
Tianhe Yu
Aviral Kumar
Rafael Rafailov
Aravind Rajeswaran
Sergey Levine
Chelsea Finn
OffRL
212
413
0
16 Feb 2021
EMaQ: Expected-Max Q-Learning Operator for Simple Yet Effective Offline
  and Online RL
EMaQ: Expected-Max Q-Learning Operator for Simple Yet Effective Offline and Online RL
Seyed Kamyar Seyed Ghasemipour
Dale Schuurmans
S. Gu
OffRL
207
119
0
21 Jul 2020
1