ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2302.08738
  4. Cited By
Exploiting Unlabeled Data for Feedback Efficient Human Preference based
  Reinforcement Learning

Exploiting Unlabeled Data for Feedback Efficient Human Preference based Reinforcement Learning

17 February 2023
Mudit Verma
Siddhant Bhambri
Subbarao Kambhampati
ArXivPDFHTML

Papers citing "Exploiting Unlabeled Data for Feedback Efficient Human Preference based Reinforcement Learning"

4 / 4 papers shown
Title
Multi-Type Preference Learning: Empowering Preference-Based
  Reinforcement Learning with Equal Preferences
Multi-Type Preference Learning: Empowering Preference-Based Reinforcement Learning with Equal Preferences
Z. Liu
Junjie Xu
Xingjiao Wu
J. Yang
Liang He
26
0
0
11 Sep 2024
Leveraging Sub-Optimal Data for Human-in-the-Loop Reinforcement Learning
Leveraging Sub-Optimal Data for Human-in-the-Loop Reinforcement Learning
Calarina Muslimani
M. E. Taylor
OffRL
38
2
0
30 Apr 2024
Methods and Mechanisms for Interactive Novelty Handling in Adversarial
  Environments
Methods and Mechanisms for Interactive Novelty Handling in Adversarial Environments
Tung Thai
Mingyu Shen
M. Garg
Ayush Kalani
Nakul Vaidya
...
Neeraj Varshney
Chitta Baral
Subbarao Kambhampati
Jivko Sinapov
matthias. scheutz
17
0
0
28 Feb 2023
Bridging the Gap: Providing Post-Hoc Symbolic Explanations for
  Sequential Decision-Making Problems with Inscrutable Representations
Bridging the Gap: Providing Post-Hoc Symbolic Explanations for Sequential Decision-Making Problems with Inscrutable Representations
S. Sreedharan
Utkarsh Soni
Mudit Verma
Siddharth Srivastava
S. Kambhampati
63
30
0
04 Feb 2020
1