Redirecting to original paper in 30 seconds...
Click below to go immediately or wait for automatic redirect
📄 Abstract
Abstract: The rapid growth of deep learning has brought about powerful models that can
handle various tasks, like identifying images and understanding language.
However, adversarial attacks, an unnoticed alteration, can deceive models,
leading to inaccurate predictions. In this paper, a generative adversarial
attack method is proposed that uses the CLIP model to create highly effective
and visually imperceptible adversarial perturbations. The CLIP model's ability
to align text and image representation helps incorporate natural language
semantics with a guided loss to generate effective adversarial examples that
look identical to the original inputs. This integration allows extensive scene
manipulation, creating perturbations in multi-object environments specifically
designed to deceive multilabel classifiers. Our approach integrates the
concentrated perturbation strategy from Saliency-based Auto-Encoder (SSAE) with
the dissimilar text embeddings similar to Generative Adversarial Multi-Object
Scene Attacks (GAMA), resulting in perturbations that both deceive
classification models and maintain high structural similarity to the original
images. The model was tested on various tasks across diverse black-box victim
models. The experimental results show that our method performs competitively,
achieving comparable or superior results to existing techniques, while
preserving greater visual fidelity.
Authors (4)
Sampriti Soor
Alik Pramanick
Jothiprakash K
Arijit Sur
Submitted
November 3, 2025
Key Contributions
Proposes a generative adversarial attack method that leverages the CLIP model to create highly effective and visually imperceptible adversarial perturbations. By incorporating natural language semantics via CLIP, the attack can perform extensive scene manipulation, specifically designed to deceive multilabel classifiers in multi-object environments.
Business Value
Helps in understanding and improving the security and robustness of AI systems, particularly those dealing with visual and textual information, by identifying vulnerabilities.