DeceptPrompt: Exploiting LLM-driven Code Generation via Adversarial Natural Language Instructions
PreviousForcing Generative Models to Degenerate Ones: The Power of Data Poisoning AttacksNextAnalyzing the Inherent Response Tendency of LLMs: Real-World Instructions-Driven Jailbreak
Last updated

