ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2312.09337
  4. Cited By
Promptable Behaviors: Personalizing Multi-Objective Rewards from Human
  Preferences

Promptable Behaviors: Personalizing Multi-Objective Rewards from Human Preferences

14 December 2023
Minyoung Hwang
Luca Weihs
Chanwoo Park
Kimin Lee
Aniruddha Kembhavi
Kiana Ehsani
ArXivPDFHTML

Papers citing "Promptable Behaviors: Personalizing Multi-Objective Rewards from Human Preferences"

15 / 15 papers shown
Title
Optimal Interactive Learning on the Job via Facility Location Planning
Optimal Interactive Learning on the Job via Facility Location Planning
Shivam Vats
Michelle Zhao
Patrick Callaghan
Mingxi Jia
Maxim Likhachev
Oliver Kroemer
George Konidaris
29
0
0
01 May 2025
Incentivizing Multi-Tenant Split Federated Learning for Foundation Models at the Network Edge
Songyuan Li
Jia Hu
Geyong Min
Haojun Huang
FedML
55
0
0
06 Mar 2025
Personalized Generation In Large Model Era: A Survey
Yiyan Xu
Jinghao Zhang
Alireza Salemi
Xinting Hu
W. Wang
Fuli Feng
Hamed Zamani
Xiangnan He
Tat-Seng Chua
3DV
71
2
0
04 Mar 2025
Robust Multi-Objective Preference Alignment with Online DPO
Raghav Gupta
Ryan Sullivan
Yunxuan Li
Samrat Phatale
Abhinav Rastogi
27
0
0
01 Mar 2025
Active Preference-based Learning for Multi-dimensional Personalization
Active Preference-based Learning for Multi-dimensional Personalization
Minhyeon Oh
Seungjoon Lee
Jungseul Ok
18
1
0
01 Nov 2024
How to Find the Exact Pareto Front for Multi-Objective MDPs?
How to Find the Exact Pareto Front for Multi-Objective MDPs?
Yining Li
Peizhong Ju
Ness B. Shroff
32
0
0
21 Oct 2024
Inferring Preferences from Demonstrations in Multi-objective
  Reinforcement Learning
Inferring Preferences from Demonstrations in Multi-objective Reinforcement Learning
Junlin Lu
Patrick Mannion
Karl Mason
18
1
0
30 Sep 2024
MotIF: Motion Instruction Fine-tuning
MotIF: Motion Instruction Fine-tuning
Minyoung Hwang
Joey Hejna
Dorsa Sadigh
Yonatan Bisk
31
1
0
16 Sep 2024
Two Tales of Persona in LLMs: A Survey of Role-Playing and
  Personalization
Two Tales of Persona in LLMs: A Survey of Role-Playing and Personalization
Yu-Min Tseng
Yu-Chao Huang
Teng-Yun Hsiao
Yu-Ching Hsu
Chao-Wei Huang
Jia-Yin Foo
Yun-Nung Chen
LLMAG
243
63
0
03 Jun 2024
RLHF from Heterogeneous Feedback via Personalization and Preference
  Aggregation
RLHF from Heterogeneous Feedback via Personalization and Preference Aggregation
Chanwoo Park
Mingyang Liu
Dingwen Kong
Kaiqing Zhang
Asuman Ozdaglar
18
28
0
30 Apr 2024
Leveraging Sub-Optimal Data for Human-in-the-Loop Reinforcement Learning
Leveraging Sub-Optimal Data for Human-in-the-Loop Reinforcement Learning
Calarina Muslimani
M. E. Taylor
OffRL
38
2
0
30 Apr 2024
Panacea: Pareto Alignment via Preference Adaptation for LLMs
Panacea: Pareto Alignment via Preference Adaptation for LLMs
Yifan Zhong
Chengdong Ma
Xiaoyuan Zhang
Ziran Yang
Haojun Chen
Qingfu Zhang
Siyuan Qi
Yaodong Yang
41
30
0
03 Feb 2024
Selective Visual Representations Improve Convergence and Generalization
  for Embodied AI
Selective Visual Representations Improve Convergence and Generalization for Embodied AI
Ainaz Eftekhar
Kuo-Hao Zeng
Jiafei Duan
Ali Farhadi
Aniruddha Kembhavi
Ranjay Krishna
27
13
0
07 Nov 2023
Large Language Models are Zero-Shot Reasoners
Large Language Models are Zero-Shot Reasoners
Takeshi Kojima
S. Gu
Machel Reid
Yutaka Matsuo
Yusuke Iwasawa
ReLM
LRM
291
2,712
0
24 May 2022
Training language models to follow instructions with human feedback
Training language models to follow instructions with human feedback
Long Ouyang
Jeff Wu
Xu Jiang
Diogo Almeida
Carroll L. Wainwright
...
Amanda Askell
Peter Welinder
Paul Christiano
Jan Leike
Ryan J. Lowe
OSLM
ALM
301
11,730
0
04 Mar 2022
1