notesum.ai

Published at November 10

SequentialBreak: Large Language Models Can be Fooled by Embedding Jailbreak Prompts into Sequential Prompt Chains

cs.CR
cs.AI
cs.CL
cs.LG

Released Date: November 10, 2024

Authors: Bijoy Ahmed Saiem1, MD Sadik Hossain Shanto1, Rakib Ahsan1, Md Rafi ur Rashid2

Aff.: 1Bangladesh University of Engineering and Technology, Dhaka, Bangladesh; 2Pennsylvania State University, PA, USA

Arxiv: http://arxiv.org/abs/2411.06426v1