S-Eval: Automatic and Adaptive Test Generation for Benchmarking Safety Evaluation of Large Language
PreviousHarmBench: A Standardized Evaluation Framework for Automated Red Teaming and Robust RefusalNextUnsafeBench: Benchmarking Image Safety Classifiers on Real-World and AI-Generated Images
Last updated
