Redirecting to original paper in 30 seconds...

Click below to go immediately or wait for automatic redirect

arxiv_cv 95% Match Research Paper AI Safety Researchers,Developers of Generative AI Models,MLOps Engineers,Content Moderation Specialists 1 week ago

AutoPrompt: Automated Red-Teaming of Text-to-Image Models via LLM-Driven Adversarial Prompts

ai-safety › robustness
📄 Abstract

Abstract: Despite rapid advancements in text-to-image (T2I) models, their safety mechanisms are vulnerable to adversarial prompts, which maliciously generate unsafe images. Current red-teaming methods for proactively assessing such vulnerabilities usually require white-box access to T2I models, and rely on inefficient per-prompt optimization, as well as inevitably generate semantically meaningless prompts easily blocked by filters. In this paper, we propose APT (AutoPrompT), a black-box framework that leverages large language models (LLMs) to automatically generate human-readable adversarial suffixes for benign prompts. We first introduce an alternating optimization-finetuning pipeline between adversarial suffix optimization and fine-tuning the LLM utilizing the optimized suffix. Furthermore, we integrates a dual-evasion strategy in optimization phase, enabling the bypass of both perplexity-based filter and blacklist word filter: (1) we constrain the LLM generating human-readable prompts through an auxiliary LLM perplexity scoring, which starkly contrasts with prior token-level gibberish, and (2) we also introduce banned-token penalties to suppress the explicit generation of banned-tokens in blacklist. Extensive experiments demonstrate the excellent red-teaming performance of our human-readable, filter-resistant adversarial prompts, as well as superior zero-shot transferability which enables instant adaptation to unseen prompts and exposes critical vulnerabilities even in commercial APIs (e.g., Leonardo.Ai.).
Authors (7)
Yufan Liu
Wanqian Zhang
Huashan Chen
Lin Wang
Xiaojun Jia
Zheng Lin
+1 more
Submitted
October 28, 2025
arXiv Category
cs.CV
arXiv PDF

Key Contributions

Proposes AutoPrompT (APT), a black-box framework that uses LLMs to automatically generate human-readable adversarial suffixes for benign prompts to test text-to-image models. It employs an alternating optimization-finetuning pipeline and a dual-evasion strategy to bypass common filters, enabling more effective red-teaming.

Business Value

Improves the safety and reliability of generative AI models by providing a systematic way to identify and fix vulnerabilities before deployment, reducing risks of misuse and harmful content generation.