Redirecting to original paper in 30 seconds...

Click below to go immediately or wait for automatic redirect

arxiv_cv 90% Match Research Paper AI security researchers,Machine learning engineers,Cybersecurity professionals,Developers of robust AI systems 1 day ago

A Generative Adversarial Approach to Adversarial Attacks Guided by Contrastive Language-Image Pre-trained Model

ai-safety › robustness
📄 Abstract

Abstract: The rapid growth of deep learning has brought about powerful models that can handle various tasks, like identifying images and understanding language. However, adversarial attacks, an unnoticed alteration, can deceive models, leading to inaccurate predictions. In this paper, a generative adversarial attack method is proposed that uses the CLIP model to create highly effective and visually imperceptible adversarial perturbations. The CLIP model's ability to align text and image representation helps incorporate natural language semantics with a guided loss to generate effective adversarial examples that look identical to the original inputs. This integration allows extensive scene manipulation, creating perturbations in multi-object environments specifically designed to deceive multilabel classifiers. Our approach integrates the concentrated perturbation strategy from Saliency-based Auto-Encoder (SSAE) with the dissimilar text embeddings similar to Generative Adversarial Multi-Object Scene Attacks (GAMA), resulting in perturbations that both deceive classification models and maintain high structural similarity to the original images. The model was tested on various tasks across diverse black-box victim models. The experimental results show that our method performs competitively, achieving comparable or superior results to existing techniques, while preserving greater visual fidelity.
Authors (4)
Sampriti Soor
Alik Pramanick
Jothiprakash K
Arijit Sur
Submitted
November 3, 2025
arXiv Category
cs.CV
arXiv PDF

Key Contributions

Proposes a generative adversarial attack method that leverages the CLIP model to create highly effective and visually imperceptible adversarial perturbations. By incorporating natural language semantics via CLIP, the attack can perform extensive scene manipulation, specifically designed to deceive multilabel classifiers in multi-object environments.

Business Value

Helps in understanding and improving the security and robustness of AI systems, particularly those dealing with visual and textual information, by identifying vulnerabilities.