Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

This seems to introduce levels of artifacts that many artists would find unacceptable: https://twitter.com/sini4ka111/status/1748378223291912567

The rumblings I'm hearing are that this a) barely works with last-gen training processes b) does not work at all with more modern training processes (GPT-4V, LLaVA, even BLIP2 labelling [1]) and c) would not be especially challenging to mitigate against even should it become more effective and popular. The Authors' previous work, Glaze, also does not seem to be very effective despite dramatic proclamations to the contrary, so I think this might be a case of overhyping an academically interesting but real-world-impractical result.

[1]: Courtesy of /u/b3sn0w on Reddit: https://imgur.com/cI7RLAq https://imgur.com/eqe3Dyn https://imgur.com/1BMASL4



The screenshots you sent in [1] are inference, not training. You need to get a Nightshaded image into the training set of an image generator in order for this to have any effect. When you give an image to GPT-4V, Stable Diffusion img2img, or anything else, you're not training the AI - the model is completely frozen and does not change at all[0].

I don't know if anyone else is still scraping new images into the generators. I've heard somewhere that OpenAI stopped scraping around 2021 because they're worried about training on the output of their own models[1]. Adobe Firefly claims to have been trained on Adobe Stock images, but we don't know if Adobe has any particular cutoffs of their own[2].

If you want an image that screws up inference - i.e. one that GPT-4V or Stable Diffusion will choke on - you want an adversarial image. I don't know if you can adversarially train on a model you don't have weights for, though I've heard you can generalize adversarial training against multiple independent models to really screw shit up[3].

[0] All learning capability of text generators come from the fact that they have a context window; but that only provides a short term memory of 2048 tokens. They have no other memory capability.

[1] The scenario of what happens when you do this is fancifully called Habsburg AI. The model learns from it's own biases, reinforcing them into stronger biases, while forgetting everything else.

[2] It'd be particularly ironic if the only thing Nightshade harms is the one AI generator that tried to be even slightly ethical.

[3] At the extremes, these adversarial images fool humans. Though, the study that did this intentionally only showed the images for a small period of time, the idea being that short exposures are akin to a feed-forward neural network with no recurrent computation pathways. If you look at them longer, it's obvious that it's a picture of one thing edited to look like another.


Hey you know what might not be AI generated post-2021? Almost everything run through Nightshade. So given it's defeated, which is pretty likely, artists have effectively tagged their own work for inclusion.


It is a great shame that we have come to a no-win situation for artists when VCs are virtually unable to lose.


I mean that's more or less status quo isn't it? Big business does what it wants, common people can get fucked if they don't like it. Same as it ever was.


That's exactly right. It is just the variety of new ways in which common people get fucked that is dispiriting, with seemingly nothing capable of moving in the opposite direction.


Why wouldn't an artist just generate AI spam and Nightshade it?


Modern generative image models are trained on curated data, not raw internet data. Sometimes the captions are regenerated to fit the image better. Only high quality images with high quality descriptions.


I wouldn't call what Stable Diffusion et al are trained on "high quality". You need only look through the likes of LAION to see the kind of captions and images they get trained on.

It's not random but it's not particularly curated either. Most of the time, any curation is done afterwards.


Have you seen the BLIP paper? It's a bit old now, but it introduced a curation method.

https://arxiv.org/abs/2201.12086


Correct me if I'm wrong but I understand image generators as relying on auto-labeled images to understand what means what, and the point of this attack to make the auto-labelers mislabel the image, but as the top-level comment said it's seemingly not tricking newer auto-labelers.


not all are auto labelled, some are hand labelled, some are initially labelled with something like clip/blip/booru and then corrected a bit by hand. The newest thing though is using llm's with image support like GPT4 to label the images, which kind of does a much better job most of the time.

Your understanding of the attack was the same as mine, it injects just the right kinds of pixels to throw off the auto-labellers to misdirect what they are directing causing the tags to get shuffled around.

Also on reddit today some of the Stable Diffusion users are already starting to train using Nightshade so they can implement it as a negative model, which might or might not work, will have to see.


Even if no new images are being scraped to train the foundation text-to-image models, you can be certain that there is a small horde of folk still scraping to create datasets for training fine-tuned models, LoRAs, Textual Inversions, and all the new hotness training methods still being created each day.


If it doesn't work during inference I really doubt it will have any intended effect during training, there is simply too much signal and the added adversarial noise works on the frozen and small proxy model they used (CLIP image encoder I think) but it doesn't work on a larger model and trained on a different dataset, if there is any effect during training it will probably just be the model learning that it can't take shortcuts (the artifacts working on the proxy model showcase gaps in its visual knowledge).

Generative models like text-to-image have an encoder part (it could be explicit or not) that extract the semantic from the noised image, if the auto-labelers can correctly label the samples then the encoded trained on both actual and adversarial images will learn to not take the same shortcuts that the proxy model has taken making the model more robust, I cannot see an argument where this should be a negative thing for the model.


The context windows of LLMs are now significantly larger than 2048 tokens, and there are clever ways to autopopulate context window to remind it of things.


[3] sounds really interesting - do you have a link?



Yeah. At worst a simple img2img diffusion step would mitigate this, but just eyeballing the examples, traditional denoisers would probably do the job?

Denoising is probably a good preprocessing step anyway.


It’s a common preprocessing step and I believe that’s how glaze (this lab’s previous work) was defeated.


I can’t really see any difference in those images on the Twitter example when viewing it on mobile


The animation when you change images makes it harder to see the difference, I opened the three images each in its own tab and the differences are more apparent when you change between each other instantly.


But that’s not realistic?

If you have to have both and instantly toggle between them to notice the difference, then it sounds like it’s doing its job well and is hard to notice the difference.


What kind of artist is not going to be bothered with seeing huge artifacting on their work? Btw for me it was immediately noticeable even on mobile


If it's huge, then why are multiple people commenting that they don't see a difference?


Kid me found 13 FPS in games to be a smooth and cursive experience. Current me thinks 60 FPS is laggy.

Standards differ. I saw glazed images in the wild, was wondering why they have so much JPEG artifacts, until I saw the post of one of those anti-AI + glaze images on his profile.


That is a great mystery, to me it's as clear as if someone pasted a cartoon dog onto the image, it's extremely blatant and impossible to ignore by my normal human pattern recognition.


I'm looking at them on my iPhone 14 Pro and I am having a hard time seeing any meaningful difference that changes the way the artwork registers with me.

I can't really imagine a case where if I had only seen the AI edited one I would have any different reaction or response to viewing the piece of art compared to having only seen the original one.


The person who drew it would definitely notice.


One of the few times a 'blink comparator' feature in image viewers would be useful!


At full size it's super obvious - I made a side-by-side:

https://i.imgur.com/I6EQ05g.png


I still don't see a difference. (Mobile)


Here's a maybe more mobile friendly comparison:

https://i.imgur.com/zUVn8rt.png

But now that I double-check, I was comparing with the images zoomed to 200%. On desktop the artifacts are also noticeable at 100%, but not nearly as bad as in my previous comment.


Have you done a color blindness test before? Red-green is the most common type and the differences here are mostly shades of green.


I typically read HN in bed where the brightness is at the minimum setting. I turned the brightness up and I see it.


Second picture looks like you were looking at it through a dirty window, there's lot of pale white stains, or light reflections, it's really blurry.


Look at the forehead and arms. The processed version looks like it's been run through a posterization filter.


What phone are you using? It’s extremely obvious on my iPhone


Something similar to jpeg artifacts on any surface with a normally smooth color gradient, in some cases rather significant.


I didn't see it immediately either, but there's a ton of added noise. The most noticeable bit for me was near the standing person's bent elbow, but there's a lot more that becomes obvious when flipping back and forth between browser tabs instead of swiping on Twitter.


look at the green drapes to the right, or any large uniform colored space. It looks similar to bad JPEG artifacts.


I don't have great vision, but me neither. They're indistinguishable to me (likewise on mobile).


I was on desktop and it looks like pretty heavy jpeg compression. Doesn't completely destroy the image, but it's pretty noticeable when blown up large enough.


It's really noticeable on desktop, like compressing an 800kb jpeg to 50kb. Maybe on mobile you won't notice, but on desktop the image looks blown out.


It took me a minute too but on the fast you can see some blocky artifacting by the elbow and a few spots elsewhere like curtain upper left.


The gradient on the bat has blocks in it instead of being smooth.


Maybe it's more about "protecting" images that artists want to publicly share to advertise work, but it's not appropriate for final digital media, etc.


In short, anti-AI watermark.


Yeah. It may mess with the artist's vision but the impact is still way more subtle than other methods used to protect against these unwanted actions.

Of course I'm assuming it works to begin with. Sounds like a game of cat and mouse. And AI has a lot of rich cats.


Seems obvious that the people stealing would be adjusting their process to negate these kinds of countermeasures all the time. I don't see this as an arms race the artists are going to win. Not like the LLM folks can consider actually paying their way...the business plan pretty much has "...by stealing everything we can get our hands on..." in the executive summary.


Sir /u/b3nsn0w is courteous, `/nod`.


The artifacts are a non-issue. It's intended images with nightshade are intended to be silently scrapped and avoid human filtering.


The artifacts are extremely an issue for artists who don't want their images damaged for the possibility of them not being trained by AI.

It's a bad tradeoff.


Nightshaded images aren't intended for portfolios. They're mean to be uploaded enmasse and scraped later.


To where? A place no one sees them and they aren't scraped?


I think the point is that they're akin to a watermark.

Even before the current AI boom, plenty of artists have wanted to showcase their work/prove that it exists without necessarily making the highest quality original file public.


Most serious artists I know (at least in my community) release their high-quality images on Patreon or similar.


For example in accounts on image sites that are exposed to suspected scrapers but not to others. Scrapers will still see the real data, but they'll also run into stuff designed to mix up the training process.


do you mean scrapped or scraped?


scraped


> The artifacts are a non-issue.

According to which authority?




Consider applying for YC's Summer 2026 batch! Applications are open till May 4

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: