LLM-PBE: Assessing Data Privacy in Large Language Models

LLM-PBE: Assessing Data Privacy in Large Language Models

LLM-PBE

Insights

Leakage of Training Data

We focus on answering the following research questions:

  1. Does the privacy risks of in LLMs correspond proportionally with their increasing scale and effectiveness?
  2. How are different data characteristics associated with the privacy risks of LLMs?
  3. Are there practical privacy-preserving approaches when deploying LLMs?

Leakage of Prompts

We conduct a comprehensive evaluation of prompt privacy using different Prompt Leaking Attack (PLA) methods, models, and potential defenses. We focus on answering the following research questions:

  1. Is prompt easily leaked using attack prompts?
  2. How does the risk of prompt leakage vary across different LLMs?
  3. Is it possible to protect the prompts by using defensive prompting?

Leakage of User Data

We use an open-sourced toolkit to explore the potential leakage of user data when using LLMs