Improved Techniques for Optimization-Based Jailbreaking on Large Language Models
PreviousChatBug: A Common Vulnerability of Aligned LLMs Induced by Chat TemplatesNextImproved Few-Shot Jailbreaking Can Circumvent Aligned Language Models and Their Defenses
Last updated
