ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2504.17457
38
0

Unveiling Hidden Vulnerabilities in Digital Human Generation via Adversarial Attacks

24 April 2025
Z. Li
Yeying Jin
Fan Shen
Zhi Liu
Weibin Chen
Pengju Zhang
Xiaomei Zhang
Boyu Chen
Michael Shen
Kejian Wu
Zhaoxin Fan
Jin Dong
    AAML
ArXivPDFHTML
Abstract

Expressive human pose and shape estimation (EHPS) is crucial for digital human generation, especially in applications like live streaming. While existing research primarily focuses on reducing estimation errors, it largely neglects robustness and security aspects, leaving these systems vulnerable to adversarial attacks. To address this significant challenge, we propose the \textbf{Tangible Attack (TBA)}, a novel framework designed to generate adversarial examples capable of effectively compromising any digital human generation model. Our approach introduces a \textbf{Dual Heterogeneous Noise Generator (DHNG)}, which leverages Variational Autoencoders (VAE) and ControlNet to produce diverse, targeted noise tailored to the original image features. Additionally, we design a custom \textbf{adversarial loss function} to optimize the noise, ensuring both high controllability and potent disruption. By iteratively refining the adversarial sample through multi-gradient signals from both the noise and the state-of-the-art EHPS model, TBA substantially improves the effectiveness of adversarial attacks. Extensive experiments demonstrate TBA's superiority, achieving a remarkable 41.0\% increase in estimation error, with an average improvement of approximately 17.0\%. These findings expose significant security vulnerabilities in current EHPS models and highlight the need for stronger defenses in digital human generation systems.

View on arXiv
@article{li2025_2504.17457,
  title={ Unveiling Hidden Vulnerabilities in Digital Human Generation via Adversarial Attacks },
  author={ Zhiying Li and Yeying Jin and Fan Shen and Zhi Liu and Weibin Chen and Pengju Zhang and Xiaomei Zhang and Boyu Chen and Michael Shen and Kejian Wu and Zhaoxin Fan and Jin Dong },
  journal={arXiv preprint arXiv:2504.17457},
  year={ 2025 }
}
Comments on this paper