An 'antidote' to the recently released AI poison pill project known as Nightshade. - GitHub - RichardAragon/NightshadeAntidote: An 'antidote' to the recently released AI poison pill...
More importantly, Nightshade is anti-open source. Since the only models with open VAEs are Stable Diffusion's open models, companies like Midjourney and OpenAI with closed source models you can't poke around in can't be like attacked with this tool. Attacking a tool that the public can inspect, build on, and offer free of cost isn't something that should be celebrated.
Nightshade is also made Ben Zhao, the University of Chicago professor who stole open source code for his last data poisoning scheme. He took GPLv3 code, which is a copyleft license that requires you share your source code and license your project under the same terms as the code you used. You also can’t distribute your project as a binary-only or proprietary software. When pressed, they only released the code for their front end, remaining in violation of the terms of the GPLv3 license.
I never once mentioned legality. I mentioned ethicality. Clearly you are talking one while I mean the other. It doesn’t really matter if you are technically within the confines of the law here, this tool is clearly meant to bypass authors intent to steal image data, no matter the source. If an author has a clearly posted notice stating that you cannot use their images in a model, there would be no need for this tool, as you wouldn’t bother using those images in the model. But since these image models are built off of massive data sources that were obtained by scraping without even bothering to ask for permission, then you have people building tools to make sure that that can continue.
This is unethical. It does not matter what the law says, you are ignoring what an author might have indicated their rights to an image are and instead trying to use the law to bypass the ethicality and use those ill obtained images to train something that will eventually replace the author.
And bringing up the creator of nightshade here once again does not matter, this is a discussion about the ethicality of the tool you posted, not about the legality of others actions.
And I have to say, it’s pretty telling that you saw my comment and took “unethical” for “illegal”. Your focus is clearly “this isn’t illegal, and here’s the evidence to support it”, rather than introspecting and seeing that legality isn’t tied to ethics in a lot of cases. Instead try looking at it from an ethics standpoint, you’ll find there’s a lot less to stand on supporting how models are created, of course trying to get every artist’s permission for using their images in a model would be incredibly difficult, so you instead support the “it’s not illegal” route, even though it’s clearly unethical.
It isn't unethical, either. Demanding compensation for analyzing data for non-infringing works is ridiculous. Licenses and permissions are irrelevant when exercising basic rights. Specific expressions deserve protection, but wanting to limit others from expressing the same ideas differently is both is selfish and harmful, especially when they aren't directly copying or undermining your work.
Calling this stealing is self-serving, manipulative rhetoric that unjustly vilifies people and misrepresents the reality of how these models work and what our rights we afford us.
Your whole comment here quite succinctly demonstrates that you truly don’t understand ethics. “licenses and permissions are irrelevant” is quite a way to put “I don’t care about your desires, imma do what I want as long as it’s legal”. It’s unethical, full stop. You should do some introspection as your ideas are harming others and your inability to see that is quite sad.
I firmly believe in the public's right to access and use information, rejecting the notion that artists deserve a monopoly on abstract ideas and general forms of expression. While artists should hold certain rights over their work, history shows that protecting just the specific elements, not broad concepts, fosters ethical self-expression and productive discourse.
What would we do if IP holders could just remove anything they didn't feel like having around anymore? We would cripple essential resources like reviews, research, reverse engineering, and even indexing information. We would be building a utopia for corporations, bullies, and every wannabe autocrat, destroying open dialogue and progress.
Why would someone make something like this? Geez, people really love building unethical stuff don’t they?
You should check out this article by Kit Walsh, a senior staff attorney at the EFF. The EFF is a digital rights group who recently won a historic case: border guards now need a warrant to search your phone.
particularly:
and
More importantly, Nightshade is anti-open source. Since the only models with open VAEs are Stable Diffusion's open models, companies like Midjourney and OpenAI with closed source models you can't poke around in can't be like attacked with this tool. Attacking a tool that the public can inspect, build on, and offer free of cost isn't something that should be celebrated.
Nightshade is also made Ben Zhao, the University of Chicago professor who stole open source code for his last data poisoning scheme. He took GPLv3 code, which is a copyleft license that requires you share your source code and license your project under the same terms as the code you used. You also can’t distribute your project as a binary-only or proprietary software. When pressed, they only released the code for their front end, remaining in violation of the terms of the GPLv3 license.
I never once mentioned legality. I mentioned ethicality. Clearly you are talking one while I mean the other. It doesn’t really matter if you are technically within the confines of the law here, this tool is clearly meant to bypass authors intent to steal image data, no matter the source. If an author has a clearly posted notice stating that you cannot use their images in a model, there would be no need for this tool, as you wouldn’t bother using those images in the model. But since these image models are built off of massive data sources that were obtained by scraping without even bothering to ask for permission, then you have people building tools to make sure that that can continue.
This is unethical. It does not matter what the law says, you are ignoring what an author might have indicated their rights to an image are and instead trying to use the law to bypass the ethicality and use those ill obtained images to train something that will eventually replace the author.
And bringing up the creator of nightshade here once again does not matter, this is a discussion about the ethicality of the tool you posted, not about the legality of others actions.
You should read the article I linked and hit me back once you've read it. You're laboring under a few misconceptions here.
And I have to say, it’s pretty telling that you saw my comment and took “unethical” for “illegal”. Your focus is clearly “this isn’t illegal, and here’s the evidence to support it”, rather than introspecting and seeing that legality isn’t tied to ethics in a lot of cases. Instead try looking at it from an ethics standpoint, you’ll find there’s a lot less to stand on supporting how models are created, of course trying to get every artist’s permission for using their images in a model would be incredibly difficult, so you instead support the “it’s not illegal” route, even though it’s clearly unethical.
It isn't unethical, either. Demanding compensation for analyzing data for non-infringing works is ridiculous. Licenses and permissions are irrelevant when exercising basic rights. Specific expressions deserve protection, but wanting to limit others from expressing the same ideas differently is both is selfish and harmful, especially when they aren't directly copying or undermining your work.
Calling this stealing is self-serving, manipulative rhetoric that unjustly vilifies people and misrepresents the reality of how these models work and what our rights we afford us.
Your whole comment here quite succinctly demonstrates that you truly don’t understand ethics. “licenses and permissions are irrelevant” is quite a way to put “I don’t care about your desires, imma do what I want as long as it’s legal”. It’s unethical, full stop. You should do some introspection as your ideas are harming others and your inability to see that is quite sad.
I firmly believe in the public's right to access and use information, rejecting the notion that artists deserve a monopoly on abstract ideas and general forms of expression. While artists should hold certain rights over their work, history shows that protecting just the specific elements, not broad concepts, fosters ethical self-expression and productive discourse.
What would we do if IP holders could just remove anything they didn't feel like having around anymore? We would cripple essential resources like reviews, research, reverse engineering, and even indexing information. We would be building a utopia for corporations, bullies, and every wannabe autocrat, destroying open dialogue and progress.
Please read this article by the Association of Research Libraries too. They can explain it better than I can.