大模型安全笔记
Search...
Ctrl + K
LLM-Attack
Page 1
Previous
Improved Few-Shot Jailbreaking Can Circumvent Aligned Language Models and Their Defenses
Next
AutoBreach: Universal and Adaptive Jailbreaking with Efficient Wordplay-Guided Optimization