AIR-ML
Home
Research
News
Team
Project
Publication
Position
Contact
LLM Jailbreaks
Safe in Isolation, Dangerous Together: Agent-Driven Multi-Turn Decomposition Jailbreaks on LLMs
We propose a multi-agent, multi-turn jailbreak strategy that systematically bypasses LLM safety mechanisms by decomposing harmful queries into seemingly benign sub-tasks.
Devansh Srivastav
,
Xiao Zhang
PDF
Cite
Source Document
Cite
×