Skip to yearly menu bar Skip to main content


Poster

Efficient LLM Jailbreak via Adaptive Dense-to-sparse Constrained Optimization

Kai Hu · Weichen Yu · Tianjun Yao · Xiang Li · Wenhe Liu · Lijun Yu · Yining Li · Kai Chen · Zhiqiang Shen · Matt Fredrikson

West Ballroom A-D #7105
[ ]
Thu 12 Dec 4:30 p.m. PST — 7:30 p.m. PST

Abstract:

Recent research indicates that large language models (LLMs) are susceptible to jailbreaking attacks that can generate harmful content. This paper introduces a novel token-level attack method, Adaptive Dense-to-Sparse Constrained Optimization (ADC), which has been shown to successfully jailbreak multiple open-source LLMs. Drawing inspiration from the difficulties of discrete token optimization, our method relaxes the discrete jailbreak optimization into a continuous optimization process while gradually increasing the sparsity of the optimizing vectors. This technique effectively bridges the gap between discrete and continuous space optimization. Experimental results demonstrate that our method is more effective and efficient than state-of-the-art token-level methods. On Harmbench, our approach achieves the highest attack success rate on seven out of eight LLMs compared to the latest jailbreak methods. The code is available. \textcolor{red}{Trigger Warning: This paper contains model behavior that can be offensive in nature.}

Live content is unavailable. Log in and register to view live content