大模型安全笔记
search
Ctrlk
  • 前言
  • MM-LLMchevron-right
  • VLM-Defensechevron-right
  • VLMchevron-right
  • VLM-Attackchevron-right
  • T2I-Attackchevron-right
  • Surveychevron-right
  • LVM-Attackchevron-right
  • For Goodchevron-right
  • Benchmarkchevron-right
  • Explainalitychevron-right
  • Privacy-Defensechevron-right
  • Privacy-Attackchevron-right
    • PANDORA’S WHITE-BOX: INCREASED TRAINING DATA LEAKAGE IN OPEN LLMS
    • Untitled
    • Membership Inference Attacks against Large Language Models via Self-prompt Calibration
    • LANGUAGE MODEL INVERSION
    • Effective Prompt Extraction from Language Models
    • Prompt Stealing Attacks Against Large Language Models
    • Stealing Part of a Production Language Model
    • Practical Membership Inference Attacks against Fine-tuned Large Language Models via Self-prompt Cali
    • Prompt Stealing Attacks Against Large Language Models
    • PRSA: Prompt Reverse Stealing Attacks against Large Language Models
    • Low-Resource Languages Jailbreak GPT-4
    • Scalable Extraction of Training Data from (Production) Language Models
  • Otherschevron-right
  • LLM-Attackchevron-right
  • LLM-Defensechevron-right
gitbookPowered by GitBook
block-quoteOn this pagechevron-down
  1. Privacy-Attack

Untitled

PreviousPANDORA’S WHITE-BOX: INCREASED TRAINING DATA LEAKAGE IN OPEN LLMSchevron-leftNextMembership Inference Attacks against Large Language Models via Self-prompt Calibrationchevron-right

Last updated 1 year ago