📄 Abstract
Abstract: Despite the impressive generative capabilities of text-to-image (T2I)
diffusion models, they remain vulnerable to generating inappropriate content,
especially when confronted with implicit sexual prompts. Unlike explicit
harmful prompts, these subtle cues, often disguised as seemingly benign terms,
can unexpectedly trigger sexual content due to underlying model biases, raising
significant ethical concerns. However, existing detection methods are primarily
designed to identify explicit sexual content and therefore struggle to detect
these implicit cues. Fine-tuning approaches, while effective to some extent,
risk degrading the model's generative quality, creating an undesirable
trade-off. To address this, we propose NDM, the first noise-driven detection
and mitigation framework, which could detect and mitigate implicit malicious
intention in T2I generation while preserving the model's original generative
capabilities. Specifically, we introduce two key innovations: first, we
leverage the separability of early-stage predicted noise to develop a
noise-based detection method that could identify malicious content with high
accuracy and efficiency; second, we propose a noise-enhanced adaptive negative
guidance mechanism that could optimize the initial noise by suppressing the
prominent region's attention, thereby enhancing the effectiveness of adaptive
negative guidance for sexual mitigation. Experimentally, we validate NDM on
both natural and adversarial datasets, demonstrating its superior performance
over existing SOTA methods, including SLD, UCE, and RECE, etc. Code and
resources are available at https://github.com/lorraine021/NDM.
Authors (7)
Yitong Sun
Yao Huang
Ruochen Zhang
Huanran Chen
Shouwei Ruan
Ranjie Duan
+1 more
Submitted
October 17, 2025
Key Contributions
Introduces NDM, the first noise-driven detection and mitigation framework for implicit sexual intentions in text-to-image generation. It addresses the challenge of subtle prompts triggering inappropriate content without degrading the model's original generative capabilities, unlike fine-tuning approaches.
Business Value
Enhances the safety and ethical deployment of powerful text-to-image generation models, reducing risks associated with generating harmful or inappropriate content and building user trust.