Redirecting to original paper in 30 seconds...

Click below to go immediately or wait for automatic redirect

arxiv_ml 98% Match Research Paper AI safety researchers,LLM developers,Cybersecurity professionals,ML engineers 1 week ago

Iterative Self-Tuning LLMs for Enhanced Jailbreaking Capabilities

large-language-models › alignment
📄 Abstract

Abstract: Recent research has shown that Large Language Models (LLMs) are vulnerable to automated jailbreak attacks, where adversarial suffixes crafted by algorithms appended to harmful queries bypass safety alignment and trigger unintended responses. Current methods for generating these suffixes are computationally expensive and have low Attack Success Rates (ASR), especially against well-aligned models like Llama2 and Llama3. To overcome these limitations, we introduce ADV-LLM, an iterative self-tuning process that crafts adversarial LLMs with enhanced jailbreak ability. Our framework significantly reduces the computational cost of generating adversarial suffixes while achieving nearly 100\% ASR on various open-source LLMs. Moreover, it exhibits strong attack transferability to closed-source models, achieving 99\% ASR on GPT-3.5 and 49\% ASR on GPT-4, despite being optimized solely on Llama3. Beyond improving jailbreak ability, ADV-LLM provides valuable insights for future safety alignment research through its ability to generate large datasets for studying LLM safety. Our code is available at: https://github.com/SunChungEn/ADV-LLM
Authors (8)
Chung-En Sun
Xiaodong Liu
Weiwei Yang
Tsui-Wei Weng
Hao Cheng
Aidan San
+2 more
Submitted
October 24, 2024
arXiv Category
cs.CL
arXiv PDF

Key Contributions

ADV-LLM introduces an iterative self-tuning process to craft adversarial LLMs with significantly enhanced jailbreak capabilities. This method drastically reduces computational cost while achieving near-perfect attack success rates on various LLMs, demonstrating strong transferability even to closed-source models.

Business Value

Highlights critical vulnerabilities in LLM safety alignment, driving the development of more robust defenses and secure AI systems. Informs security professionals about potential attack vectors.