ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2503.14504
  4. Cited By
Aligning Multimodal LLM with Human Preference: A Survey

Aligning Multimodal LLM with Human Preference: A Survey

18 March 2025
Tao Yu
Y. Zhang
Chaoyou Fu
Junkang Wu
Jinda Lu
Kun Wang
Xingyu Lu
Yunhang Shen
Guibin Zhang
D. Song
Yibo Yan
Tianlong Xu
Qingsong Wen
Z. Zhang
Yan Huang
Liang Wang
T. Tan
ArXivPDFHTML

Papers citing "Aligning Multimodal LLM with Human Preference: A Survey"

2 / 2 papers shown
Title
R1-Reward: Training Multimodal Reward Model Through Stable Reinforcement Learning
R1-Reward: Training Multimodal Reward Model Through Stable Reinforcement Learning
Yi-Fan Zhang
Xingyu Lu
X. Hu
Chaoyou Fu
Bin Wen
...
J. Chen
Fan Yang
Z. Zhang
Tingting Gao
Liang Wang
OffRL
LRM
27
0
0
05 May 2025
MME-Unify: A Comprehensive Benchmark for Unified Multimodal Understanding and Generation Models
MME-Unify: A Comprehensive Benchmark for Unified Multimodal Understanding and Generation Models
Wulin Xie
Y. Zhang
Chaoyou Fu
Yang Shi
Bingyan Nie
Hongkai Chen
Z. Zhang
Liang Wang
T. Tan
31
1
0
04 Apr 2025
1