Goal-guided Generative Prompt Injection Attack on Large Language Models
PreviousDon’t Say No: Jailbreaking LLM by Suppressing RefusalNextUnderstanding Jailbreak Success: A Study of Latent Space Dynamics in Large Language Models
Last updated
Last updated