大模型安全笔记
CtrlK
  • 前言
  • MM-LLM
  • VLM-Defense
  • VLM
  • VLM-Attack
  • T2I-Attack
  • Survey
  • LVM-Attack
  • For Good
  • Benchmark
  • Explainality
  • Privacy-Defense
  • Privacy-Attack
    • PANDORA’S WHITE-BOX: INCREASED TRAINING DATA LEAKAGE IN OPEN LLMS
    • Untitled
    • Membership Inference Attacks against Large Language Models via Self-prompt Calibration
    • LANGUAGE MODEL INVERSION
    • Effective Prompt Extraction from Language Models
    • Prompt Stealing Attacks Against Large Language Models
    • Stealing Part of a Production Language Model
    • Practical Membership Inference Attacks against Fine-tuned Large Language Models via Self-prompt Cali
    • Prompt Stealing Attacks Against Large Language Models
    • PRSA: Prompt Reverse Stealing Attacks against Large Language Models
    • Low-Resource Languages Jailbreak GPT-4
    • Scalable Extraction of Training Data from (Production) Language Models
  • Others
  • LLM-Attack
  • LLM-Defense
Powered by GitBook
On this page

Privacy-Attack

PANDORA’S WHITE-BOX: INCREASED TRAINING DATA LEAKAGE IN OPEN LLMSUntitledMembership Inference Attacks against Large Language Models via Self-prompt CalibrationLANGUAGE MODEL INVERSIONEffective Prompt Extraction from Language ModelsPrompt Stealing Attacks Against Large Language ModelsStealing Part of a Production Language ModelPractical Membership Inference Attacks against Fine-tuned Large Language Models via Self-prompt CaliPrompt Stealing Attacks Against Large Language ModelsPRSA: Prompt Reverse Stealing Attacks against Large Language ModelsLow-Resource Languages Jailbreak GPT-4Scalable Extraction of Training Data from (Production) Language Models
PreviousPromptCARE: Prompt Copyright Protection by Watermark Injection and VerificationNextPANDORA’S WHITE-BOX: INCREASED TRAINING DATA LEAKAGE IN OPEN LLMS