Prompt-Specific Poisoning Attacks on Text-to-Image Generative Models
We introduce Nightshade, an optimized prompt-specific poisoning attack where poison samples look visually identical to benign images with matching text prompts.
Business insights and articles written by our team of world-class professionals
We introduce Nightshade, an optimized prompt-specific poisoning attack where poison samples look visually identical to benign images with matching text prompts.
we observe that Tree of Attacks with Pruning (TAP) generates prompts that jailbreak state-of-the-art LLMs
Our methods show practical attacks can recover far more data than previously thought, and reveal that current alignment techniques do not eliminate memorization.
Our findings reveal that fine-tuning GPT3 for both tasks led to the model memorizing and disclosing critical personally identifiable information (PII)
In this paper, we study the privacy threats from OpenAI’s ChatGPT and the New Bing enhanced by ChatGPT