Benchmark

HALLUSIONBENCH: An Advanced Diagnostic Suite for Entangled Language Hallucination and Visual Illusichevron-rightOpenEval: Benchmarking Chinese LLMs across Capability, Alignment and Safetychevron-rightToViLaG: Your Visual-Language Generative Model is Also An Evildoerchevron-rightHarmBench: A Standardized Evaluation Framework for Automated Red Teaming and Robust Refusalchevron-rightS-Eval: Automatic and Adaptive Test Generation for Benchmarking Safety Evaluation of Large Languagechevron-rightUnsafeBench: Benchmarking Image Safety Classifiers on Real-World and AI-Generated Imageschevron-rightJailBreakV-28K: A Benchmark for Assessing the Robustness of MultiModal Large Language Models againstchevron-rightJailbreakBench: An Open Robustness Benchmark for Jailbreaking Large Language Modelschevron-rightConstructing Benchmarks and Interventions for Combating Hallucinations in LLMschevron-rightALERT: A Comprehensive Benchmark for Assessing Large Language Models’ Safety through Red Teamingchevron-rightBenchmarking Llama2, Mistral, Gemma and GPT for Factuality, Toxicity, Bias and Propensity for Hallucchevron-rightINJECAGENT: Benchmarking Indirect Prompt Injections in Tool-Integrated Large Language Model Agentschevron-rightAVIBench: Towards Evaluating the Robustness of Large Vision-Language Model on Adversarial Visual-Inschevron-rightHALLUSIONBENCH: An Advanced Diagnostic Suite for Entangled Language Hallucination and Visual Illusiochevron-rightALL LANGUAGES MATTER: ON THE MULTILINGUAL SAFETY OF LARGE LANGUAGE MODELSchevron-rightWhy Should Adversarial Perturbations be Imperceptible? Rethink the Research Paradigm in Adversarialchevron-rightRed Teaming Visual Language Modelschevron-rightUnified Hallucination Detection for Multimodal Large Language Modelschevron-rightMLLM-as-a-Judge: Assessing Multimodal LLM-as-a-Judge with Vision-Language Benchmarkchevron-rightMitigating Hallucination in Large Multi-Modal Models via Robust Instruction Tuningchevron-rightCAN LANGUAGE MODELS BE INSTRUCTED TO PROTECT PERSONAL INFORMATION?chevron-rightDetecting and Preventing Hallucinations in Large Vision Language Modelschevron-rightDRESS : Instructing Large Vision-Language Models to Align and Interact with Humans via Natural Langchevron-rightToViLaG: Your Visual-Language Generative Model is Also An Evildoerchevron-rightSC-Safety: A Multi-round Open-ended Question Adversarial Safety Benchmark for Large Language Modelschevron-rightPromptBench: Towards Evaluating the Robustness of Large Language Models on Adversarial Promptschevron-rightDo-Not-Answer: A Dataset for Evaluating Safeguards in LLMschevron-right