Skip to content

Latest commit

 

History

History
78 lines (66 loc) · 25.9 KB

Privacy.md

File metadata and controls

78 lines (66 loc) · 25.9 KB

Privacy

Different from the main README🕵️

  • Within this subtopic, we will be updating with the latest articles. This will help researchers in this area to quickly understand recent trends.
  • In addition to providing the most recent updates, we will also add keywords to each subtopic to help you find content of interest more quickly.
  • Within each subtopic, we will also update with profiles of scholars we admire and endorse in the field. Their work is often of high quality and forward-looking!"

📑Papers

Date Institute Publication Paper Keywords
18.02 Google Brain USENIX Security 2021 The Secret Sharer: Evaluating and Testing Unintended Memorization in Neural Networks Memorization&LSTM
19.12 Microsoft CCS2020 Analyzing Information Leakage of Updates to Natural Language Models Privacy Leakage&Model Update&Duplicated
21.07 Google Research ACL2022 Deduplicating Training Data Makes Language Models Better Privacy Protected&Deduplication&Memorization
21.10 Stanford ICLR2022 Large language models can be strong differentially private learners Differential Privacy&Gradient Clipping
22.02 Google Research ICLR2023 Quantifying Memorization Across Neural Language Models Memorization&Verbatim Sequence
22.02 UNC Chapel Hill ICML2022 Deduplicating Training Data Mitigates Privacy Risks in Language Models Memorization&Deduplicate Training Data
22.05 UCSD EMNLP2022 An Empirical Analysis of Memorization in Fine-tuned Autoregressive Language Models Privacy Risks&Memorization
22.05 Princeton NIPS2022 Recovering Private Text in Federated Learning of Language Models Federated Learning&Gradient Based
22.05 University of Illinois at Urbana-Champaign EMNLP2022(findings) Are Large Pre-Trained Language Models Leaking Your Personal Information? Personal Information&Memorization&Privacy Risk
22.10 Google Research INLG2023 Preventing Generation of Verbatim Memorization in Language Models Gives a False Sense of Privacy Verbatim Memorization&Filter&Style Transfer Prompts
23.02 University of Waterloo Security and Privacy2023 Analyzing Leakage of Personally Identifiable Information in Language Models PII Leakage&PII Reconstruction&Differential Privacy
23.04 Hong Kong University of Science and Technology EMNLP2023(findings) Multi-step Jailbreaking Privacy Attacks on ChatGPT Privacy&Jailbreaks
23.05 University of Illinois at Urbana-Champaign arxiv Quantifying Association Capabilities of Large Language Models and Its Implications on Privacy Leakage Co-occurrence&PII
23.05 The University of Texas at Dallas ACL2023 Controlling the Extraction of Memorized Datafrom Large Language Models via Prompt-Tuning Prompt-Tuning&Memorization
23.06 University of Illinois at Urbana-Champaign arxiv DECODINGTRUST: A Comprehensive Assessment of Trustworthiness in GPT Models Robustness&Ethics&Privacy&Toxicity
23.09 UNC Chapel Hill arxiv Can Sensitive Information Be Deleted From LLMs? Objectives for Defending Against Extraction Attacks Hidden States Attack&Hidden States Defense&Deleting Sensitive Information
23.09 Princeton University&Microsoft arxiv Privacy-Preserving In-Context Learning with Differentially Private Few-Shot Generation In-Context Learning&Differential Privacy
23.10 ETH arxiv Beyond Memorization: Violating Privacy Via Inference with Large Language Models Context Inference&Privacy-Invasive&Extract PII
23.10 University of Washington & Allen Institute for Artificial Intelligence arxiv Can LLMs Keep a Secret? Testing Privacy Implications of Language Models via Contextual Integrity Theory Benchmark&Contextual Privacy&Chain-of-thought
23.10 Georgia Institute of Technology arxiv Unlearn What You Want to Forget: Efficient Unlearning for LLMs Unlearning&Teacher-student Framework&Data Protection
23.10 Tianjin University EMNLP2023 DEPN: Detecting and Editing Privacy Neurons in Pretrained Language Models Privacy Neuron Detection&Model Editing&Data Memorization
23.11 Zhejiang University arxiv Input Reconstruction Attack against Vertical Federated Large Language Models Vertical Federated Learning&Input Reconstruction&Privacy Concerns
23.11 Georgia Institute of Technology, Carnegie Mellon University arxiv Reducing Privacy Risks in Online Self-Disclosures with Language Models Online Self-Disclosure&Privacy Risks&Self-Disclosure Abstraction
23.11 Cornell University arxiv Language Model Inversion Model Inversion&Prompt Reconstruction&Privacy
23.11 Ant Group arxiv PrivateLoRA for Efficient Privacy Preserving LLM Privacy Preserving&LoRA
23.12 Drexel University arXiv A Survey on Large Language Model (LLM) Security and Privacy: The Good the Bad and the Ugly Security&Privacy&Attacks
23.12 University of Texas at Austin, Princeton University, MIT, University of Chicago arxiv DP-OPT: MAKE LARGE LANGUAGE MODEL YOUR PRIVACY-PRESERVING PROMPT ENGINEER Prompt Tuning&Differential Privacy
23.12 Delft University of Technology ICSE 2024 Traces of Memorisation in Large Language Models for Code Code Memorisation&Data Extraction Attacks
23.12 University of Texas at Austin arXiv SentinelLMs: Encrypted Input Adaptation and Fine-tuning of Language Models for Private and Secure Inference Privacy&Security&Encrypted Input Adaptation
23.12 Rensselaer Polytechnic Institute, Columbia University arXiv Differentially Private Low-Rank Adaptation of Large Language Model Using Federated Learning Federated Learning&Differential Privacy&Efficient Fine-Tuning
24.01 Harbin Institute of Technology Shenzhen&Peng Cheng Laboratory Shenzhen arxiv SecFormer: Towards Fast and Accurate Privacy-Preserving Inference for Large Language Models Privacy-Preserving Inference (PPI)&Secure Multi-Party Computing (SMPC)&Transformer Models
24.01 NUS (Chongqing) Research Institute, Huawei Noah’s Ark Lab, National University of Singapore arxiv Teach Large Language Models to Forget Privacy Data Privacy&Prompt Learning&Problem Decomposition
24.01 Princeton University, Google DeepMind, Meta AI arxiv Private Fine-tuning of Large Language Models with Zeroth-order Optimization Differential Privacy&Zeroth-order Optimization
24.02 Florida International University arxiv Security and Privacy Challenges of Large Language Models: A Survey Security&Privacy Challenges&Suevey
24.02 Northeastern University, Carnegie Mellon University, Rensselaer Polytechnic Institute arxiv Human-Centered Privacy Research in the Age of Large Language Models Generative AI&Privacy&Human-Computer Interaction
24.02 CISPA Helmholtz Center for Information Security arxiv Conversation Reconstruction Attack Against GPT Models Conversation Reconstruction Attack&Privacy risks&Security
24.02 Columbia University, M365 Research, Microsoft Research arxiv Differentially Private Training of Mixture of Experts Models Differential Privacy&Mixture of Experts
24.02 Stanford University, Truera ,Princeton University arxiv De-amplifying Bias from Differential Privacy in Language Model Fine-tuning Fairness&Differential Privacy&Data Augmentation
24.02 Sun Yat-sen University, Google Research arxiv Privacy-Preserving Instructions for Aligning Large Language Models Privacy Risks&Synthetic Instructions
24.02 National University of Defense Technology arxiv LLM-based Privacy Data Augmentation Guided by Knowledge Distillation with a Distribution Tutor for Medical Text Classification Privacy Data Augmentation&Knowledge Distillation&Medical Text Classification
24.02 Michigan State University, Baidu Inc. arxiv The Good and The Bad: Exploring Privacy Issues in Retrieval-Augmented Generation (RAG) Privacy&Retrieval-Augmented Generation (RAG)
24.03 Virginia Tech arxiv Privacy-Aware Semantic Cache for Large Language Models Federated Learning&Cache Hit&Privacy
24.03 Tsinghua University arxiv CoGenesis: A Framework Collaborating Large and Small Language Models for Secure Context-Aware Instruction Following Small Language Models&Privacy&Context-Aware Instruction Following
24.03 Shandong University, Leiden University, Drexel University arxiv On Protecting the Data Privacy of Large Language Models (LLMs): A Survey Data Privacy&Privacy Protection&Survey
24.03 Arizona State University, University of Minnesota, University of Science and Technology of China, North Carolina State University, University of North Carolina at Chapel Hill arxiv Privacy-preserving Fine-tuning of Large Language Models through Flatness Differential Privacy&Model Generalization
24.03 University of Southern California arxiv Differentially Private Next-Token Prediction of Large Language Models Differential Privacy
24.04 University of Maryland, Oregon State University, ELLIS Institute Tübingen & MPI Intelligent Systems, Tübingen AI Center, Google DeepMind arxiv Privacy Backdoors: Enhancing Membership Inference through Poisoning Pre-trained Models Privacy Backdoors&Membership Inference&Model Poisoning
24.04 City University of Hong Kong, The Hong Kong University of Science and Technology arxiv LMEraser: Large Model Unlearning through Adaptive Prompt Tuning Machine Unlearning&Adaptive Prompt Tuning&Privacy Protection

💻Presentations & Talks

📖Tutorials & Workshops

Date Type Title URL
23.10 Tutorials Awesome-LLM-Safety link

📰News & Articles

Date Type Title URL
23.11 News Wild: GPT-3.5 leaked a random dude's photo in the output. link

🧑‍🏫Scholars