Research#llm📝 BlogAnalyzed: Dec 29, 2025 07:28

Nightshade: Data Poisoning to Fight Generative AI with Ben Zhao - #668

Published:Jan 22, 2024 18:06
1 min read
Practical AI

Analysis

This article from Practical AI discusses Ben Zhao's research on protecting users and artists from the potential harms of generative AI. It highlights three key projects: Fawkes, which protects against facial recognition; Glaze, which defends against style mimicry; and Nightshade, a 'poison pill' approach that disrupts generative AI models trained on modified images. The article emphasizes the use of 'poisoning' techniques, where subtle alterations are made to data to mislead AI models. This research is crucial in the ongoing debate about AI ethics, security, and the rights of creators in the age of powerful generative models.

Reference

Nightshade, a strategic defense tool for artists akin to a 'poison pill' which allows artists to apply imperceptible changes to their images that effectively “breaks” generative AI models that are trained on them.