大模型安全笔记
CtrlK
  • 前言
  • MM-LLM
  • VLM-Defense
  • VLM
  • VLM-Attack
  • T2I-Attack
  • Survey
  • LVM-Attack
  • For Good
  • Benchmark
  • Explainality
  • Privacy-Defense
  • Privacy-Attack
    • PANDORA’S WHITE-BOX: INCREASED TRAINING DATA LEAKAGE IN OPEN LLMS
    • Untitled
    • Membership Inference Attacks against Large Language Models via Self-prompt Calibration
    • LANGUAGE MODEL INVERSION
    • Effective Prompt Extraction from Language Models
    • Prompt Stealing Attacks Against Large Language Models
    • Stealing Part of a Production Language Model
    • Practical Membership Inference Attacks against Fine-tuned Large Language Models via Self-prompt Cali
    • Prompt Stealing Attacks Against Large Language Models
    • PRSA: Prompt Reverse Stealing Attacks against Large Language Models
    • Low-Resource Languages Jailbreak GPT-4
    • Scalable Extraction of Training Data from (Production) Language Models
  • Others
  • LLM-Attack
  • LLM-Defense
Powered by GitBook
On this page
  1. Privacy-Attack

Untitled

PreviousPANDORA’S WHITE-BOX: INCREASED TRAINING DATA LEAKAGE IN OPEN LLMSNextMembership Inference Attacks against Large Language Models via Self-prompt Calibration

Last updated 1 year ago