Xiao Zhang
Xiao Zhang
About
Research
Publication
Teaching
Service
Contact
Light
Dark
Automatic
Large Language Models (LLMs)
Jul 29, 2025
Safe in Isolation, Dangerous Together: Agent-Driven Multi-Turn Decomposition Jailbreaks on LLMs
We propose a multi-agent, multi-turn jailbreak strategy that systematically bypasses LLM safety mechanisms by decomposing harmful queries into seemingly benign sub-tasks.
Devansh Srivastav
,
Xiao Zhang
PDF
Cite
Source Document
Cite
×