ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2502.08556
100
2

Human-Centric Foundation Models: Perception, Generation and Agentic Modeling

12 February 2025
Shixiang Tang
Y. Wang
Lu Chen
Yuan Wang
Sida Peng
Dan Xu
W. Ouyang
    VGen
ArXivPDFHTML
Abstract

Human understanding and generation are critical for modeling digital humans and humanoid embodiments. Recently, Human-centric Foundation Models (HcFMs) inspired by the success of generalist models, such as large language and vision models, have emerged to unify diverse human-centric tasks into a single framework, surpassing traditional task-specific approaches. In this survey, we present a comprehensive overview of HcFMs by proposing a taxonomy that categorizes current approaches into four groups: (1) Human-centric Perception Foundation Models that capture fine-grained features for multi-modal 2D and 3D understanding. (2) Human-centric AIGC Foundation Models that generate high-fidelity, diverse human-related content. (3) Unified Perception and Generation Models that integrate these capabilities to enhance both human understanding and synthesis. (4) Human-centric Agentic Foundation Models that extend beyond perception and generation to learn human-like intelligence and interactive behaviors for humanoid embodied tasks. We review state-of-the-art techniques, discuss emerging challenges and future research directions. This survey aims to serve as a roadmap for researchers and practitioners working towards more robust, versatile, and intelligent digital human and embodiments modeling.

View on arXiv
@article{tang2025_2502.08556,
  title={ Human-Centric Foundation Models: Perception, Generation and Agentic Modeling },
  author={ Shixiang Tang and Yizhou Wang and Lu Chen and Yuan Wang and Sida Peng and Dan Xu and Wanli Ouyang },
  journal={arXiv preprint arXiv:2502.08556},
  year={ 2025 }
}
Comments on this paper