I've been following something that I hope most of you (and especially people like
Bill Cipher) will appreciate quite a lot. It's called Nightshade, and it's an optimized prompt-specific data model poisoning attack. The full paper on it can be found here -
Prompt-Specific Poisoning Attacks on Text-to-Image Generative ModelsOne of the masterminds behind this technology is
Ben Y. Zhao who seems to have authored quite a few research papers fighting against the rise of these ML algorithms that have been poisoning the online art community. He's also developed a tool called
Glaze whose purpose is aiding artists in the fight against products like Stable Diffusion.
The way technologies like Nightshade and Glaze work is by changing the pixels of images in subtle ways that are invisible to the human eye but manipulate machine-learning models to interpret the image as something different from what it actually shows. Artists can use these tools to parse their finished art through, and have the modifications automatically applied in a way that would disrupt the learning capabilities of these models. Attached you will find a couple of images that depict the effects of those poisoned samples on the models' performance.
Being a software engineer myself, I very much respect the fact that the companies developing these technologies are making them open-source, allowing the community to modify the code and customize it to specific edge cases, as well as develop it beyond the original creator's vision and reach. Models like LAION, the one used in Stable Diffusion, contain billions of images scraped from the internet. If we can hope for any meaningful effect, we need as many poisoned samples as possible. Unfortunately, affecting the images already used in training those models is not an option, but having those technologies available will decrease the impact of these models in a (hopefully) noticeable way in the future. At this point we will take anything we have at our disposal in what feels like ultimately a losing battle against the commercialization of one of humanity's most valuable peculiarities.
The poisoned data is very difficult to remove, as it requires tech companies to painstakingly find and delete each corrupted sample, which is just unrealistic consindering the enormous amounts of data these models work with. And implementing an automated algorithm that scrapes these databases and tries to detect entries that have been tampered with is also unrealistic. It's similar to asking a person struggling with severe psychosis and schizophrenia to describe what is real and what isn't. In the end you'll get a dataset that you can't really rely on.
Generative AI models are excellent at making connections between words, which helps the poison spread. Technologies like Nightshade infect not only the word “dog” but all similar concepts, such as “puppy,” “husky,” and “wolf.” The poison attack also works on tangentially related images. For example, if the model scraped a poisoned image for the prompt “fantasy art,” the prompts “dragon” and “a castle in The Lord of the Rings” would similarly be manipulated into something else.
Naturally, people like
Vitaly Shmatikov, a professor at Cornell University who studies AI model security, are opposing this type of technologies and are urging for focus on defense against them, but that's to be expected.
And on the other side (thankfully) we have people like
Junfeng Yang, a computer science professor at Columbia University, and
Eva Toorenent, an illustrator and artist who uses Glaze, and
Autumn Beverly, another artist that has been using these tools, all say the presence of this type of technology is a step in the right direction:
Quote:It is going to make [AI companies] think twice, because they have the possibility of destroying their entire model by taking our work without our consent.
Real arists who have previously removed their work from the internet are bringing it back, but parsed through tools like Nightshade and Glaze, and say it has brought their confidence and sense of safety back, since models like LAION won't be able to unconsensually scrape their work and provide it to others for free.
We have yet to see the extent of these tools' impact on ML models that generate "art", but I believe this is a step in the right direction as well. Artists should not have to fear posting their work online. Art is a fundamental expression of human creativity, and a materialization of the human soul. Commercializing that and making it nothing more than a commodity that fuels the engine of capitalism should be considered a crime against humanity.
Let me know your thoughts on this.
Love & Light
Nydex attached the following image(s):
1.png
(2,210kb) downloaded 178 time(s). 2.png
(742kb) downloaded 174 time(s).