Hidden You Malicious Goal Into Benign Narratives: Jailbreak Large Language Models through Logic Chai
PreviousSHADOW ALIGNMENT: THE EASE OF SUBVERTING SAFELY-ALIGNED LANGUAGE MODELSNextJailbreakLens: Visual Analysis of Jailbreak Attacks Against Large Language Models
Last updated
