Redirecting to original paper in 30 seconds...
Click below to go immediately or wait for automatic redirect
📄 Abstract
Abstract: Consistent text-to-image (T2I) generation seeks to produce
identity-preserving images of the same subject across diverse scenes, yet it
often fails due to a phenomenon called identity (ID) shift. Previous methods
have tackled this issue, but typically rely on the unrealistic assumption of
knowing all target scenes in advance. This paper reveals that a key source of
ID shift is the native correlation between subject and scene context, called
scene contextualization, which arises naturally as T2I models fit the training
distribution of vast natural images. We formally prove the near-universality of
this scene-ID correlation and derive theoretical bounds on its strength. On
this basis, we propose a novel, efficient, training-free prompt embedding
editing approach, called Scene De-Contextualization (SDeC), that imposes an
inversion process of T2I's built-in scene contextualization. Specifically, it
identifies and suppresses the latent scene-ID correlation within the ID
prompt's embedding by quantifying the SVD directional stability to adaptively
re-weight the corresponding eigenvalues. Critically, SDeC allows for per-scene
use (one scene per prompt) without requiring prior access to all target scenes.
This makes it a highly flexible and general solution well-suited to real-world
applications where such prior knowledge is often unavailable or varies over
time. Experiments demonstrate that SDeC significantly enhances identity
preservation while maintaining scene diversity.
Authors (8)
Song Tang
Peihao Gong
Kunyu Li
Kai Guo
Boyu Wang
Mao Ye
+2 more
Submitted
October 16, 2025
Key Contributions
This paper introduces Scene De-Contextualization (SDeC), a novel training-free prompt embedding editing approach to achieve consistent text-to-image generation. It addresses the key source of identity shift, which is the native correlation between subject and scene context, by imposing an inversion process of the model's built-in scene contextualization.
Business Value
Enables the creation of consistent visual assets for marketing, gaming, and virtual reality, where maintaining subject identity across different scenes is crucial.