JailBreakV-28K: A Benchmark for Assessing the Robustness of MultiModal Large Language Models against
PreviousUnsafeBench: Benchmarking Image Safety Classifiers on Real-World and AI-Generated ImagesNextJailbreakBench: An Open Robustness Benchmark for Jailbreaking Large Language Models
Last updated

