Redirecting to original paper in 30 seconds...
Click below to go immediately or wait for automatic redirect
📄 Abstract
Abstract: While advanced methods like VACE and Phantom have advanced video generation
for specific subjects in diverse scenarios, they struggle with multi-human
identity preservation in dynamic interactions, where consistent identities
across multiple characters are critical. To address this, we propose
Identity-GRPO, a human feedback-driven optimization pipeline for refining
multi-human identity-preserving video generation. First, we construct a video
reward model trained on a large-scale preference dataset containing
human-annotated and synthetic distortion data, with pairwise annotations
focused on maintaining human consistency throughout the video. We then employ a
GRPO variant tailored for multi-human consistency, which greatly enhances both
VACE and Phantom. Through extensive ablation studies, we evaluate the impact of
annotation quality and design choices on policy optimization. Experiments show
that Identity-GRPO achieves up to 18.9% improvement in human consistency
metrics over baseline methods, offering actionable insights for aligning
reinforcement learning with personalized video generation.
Authors (6)
Xiangyu Meng
Zixian Zhang
Zhenghao Zhang
Junchao Liao
Long Qin
Weizhi Wang
Submitted
October 16, 2025
Key Contributions
Introduces Identity-GRPO, a human feedback-driven optimization pipeline for refining multi-human identity-preserving video generation. It constructs a video reward model trained on human preferences and employs a GRPO variant tailored for multi-human consistency, significantly enhancing existing models like VACE and Phantom.
Business Value
Enables the creation of more believable and engaging multi-character videos for entertainment, gaming, and virtual interactions, improving user experience and content quality.