University of Chicago researchers have unveiled Nightshade, a tool designed to disrupt AI models attempting to learn from artistic imagery.
- 10 Best High-Pitched Devices to Annoy Neighbors Effectively
- European VC ecosystem heavily biased against female founders
- TNW Podcast: Oscar Kneppers on academic startups; ESA wants to detect space-time ripples
- AstraZeneca password lapse exposed patient data
- Coatue backs Kubecost’s goal to supply data to millions of Kubernetes developers
The tool – still in its developmental phase – allows artists to protect their work by subtly altering pixels in images, rendering them imperceptibly different to the human eye but confusing to AI models.
Many artists and creators have expressed concern over the use of their work in training commercial AI products without their consent.
AI models rely on vast amounts of multimedia data – including written material and images, often scraped from the web – to function effectively. Nightshade offers a potential solution by sabotaging this data.
When integrated into digital artwork, Nightshade misleads AI models, causing them to misidentify objects and scenes.
For instance, Nightshade transformed images of dogs into data that appeared to AI models as cats. After exposure to a mere 100 poison samples, the AI reliably generated a cat when asked for a dog—demonstrating the tool’s effectiveness.
This technique not only confuses AI models but also challenges the fundamental way in which generative AI operates. By exploiting the clustering of similar words and ideas in AI models, Nightshade can manipulate responses to specific prompts and further undermine the accuracy of AI-generated content.
Developed by computer science professor Ben Zhao and his team, Nightshade is an extension of their prior product, Glaze, which cloaks digital artwork and distorts pixels to baffle AI models regarding artistic style.
While the potential for misuse of Nightshade is acknowledged, the researchers’ primary objective is to shift the balance of power from AI companies back to artists and discourage intellectual property violations.
The introduction of Nightshade presents a major challenge to AI developers. Detecting and removing images with poisoned pixels is a complex task, given the imperceptible nature of the alterations.
As the researchers await peer review of their work, Nightshade is a beacon of hope for artists seeking to protect their creative endeavours.
See also: UMG files landmark lawsuit against AI developer Anthropic
Explore other upcoming enterprise technology events and webinars powered by TechForge here.