Content not yet available
This lecture has no active video or poster.
Would you like to see your presentation here, made available to a global audience of researchers?
Add your own presentation or have us affordably record your next conference.
We identify a jailbreaking vulnerability in multiple open-source LLMs: by augmenting dangerous requests using certain distractors" to obfuscate their intent, we elicit specific, actionable responses on a wide variety of harmful topics. We find that such an attack noticeably alters the contents of these models' chains of thought, including changed frequencies of seemingly unrelated $n$-grams and heightened ethical scrutiny about harmful requests even when their response is ultimately jailbroken.