February 2024 | BADHAN CHANDRA DAS, M. HADI AMINI, YANZHAO WU
The paper "Security and Privacy Challenges of Large Language Models: A Survey" by Badhan Chandra Das, M. Hadi Amini, and Yanzhao Wu provides a comprehensive review of the security and privacy challenges faced by Large Language Models (LLMs). LLMs, such as GPT-3 and GPT-4, have become increasingly popular in various fields due to their advanced capabilities in text generation, summarization, translation, and question-answering. However, these models are vulnerable to security and privacy attacks, including data poisoning, Personally Identifiable Information (PII) leakage, and adversarial attacks. The authors categorize these attacks into security and privacy issues, focusing on prompt hacking (jailbreaking and prompt injection), adversarial attacks (backdoor and data poisoning), and privacy attacks (gradient leakage, membership inference, and PII leakage). They also discuss existing defense mechanisms and highlight research gaps and future directions. The paper aims to provide a clear understanding of the security and privacy challenges of LLMs, contributing to the development of secure and privacy-preserving human-LLM interactions.The paper "Security and Privacy Challenges of Large Language Models: A Survey" by Badhan Chandra Das, M. Hadi Amini, and Yanzhao Wu provides a comprehensive review of the security and privacy challenges faced by Large Language Models (LLMs). LLMs, such as GPT-3 and GPT-4, have become increasingly popular in various fields due to their advanced capabilities in text generation, summarization, translation, and question-answering. However, these models are vulnerable to security and privacy attacks, including data poisoning, Personally Identifiable Information (PII) leakage, and adversarial attacks. The authors categorize these attacks into security and privacy issues, focusing on prompt hacking (jailbreaking and prompt injection), adversarial attacks (backdoor and data poisoning), and privacy attacks (gradient leakage, membership inference, and PII leakage). They also discuss existing defense mechanisms and highlight research gaps and future directions. The paper aims to provide a clear understanding of the security and privacy challenges of LLMs, contributing to the development of secure and privacy-preserving human-LLM interactions.