this post was submitted on 15 Dec 2025
659 points (98.5% liked)

Technology

77090 readers
3054 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related news or articles.
  3. Be excellent to each other!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, this includes using AI responses and summaries. To ask if your bot can be added please contact a mod.
  9. Check for duplicates before posting, duplicates may be removed
  10. Accounts 7 days and younger will have their posts automatically removed.

Approved Bots


founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] PrivateNoob@sopuli.xyz 39 points 1 day ago* (last edited 1 day ago) (5 children)

There are poisoning scripts for images, where some random pixels have totally nonsensical / erratic colors, which we won't really notice at all, however this would wreck the LLM into shambles.

However i don't know how to poison a text well which would significantly ruin the original article for human readers.

Ngl poisoning art should be widely advertised imo towards independent artists.

[–] turdas@suppo.fi 25 points 1 day ago (1 children)

The I in LLM stands for "image".

[–] PrivateNoob@sopuli.xyz 7 points 1 day ago

Fair enough on the technicality issues, but you get my point. I think just some art poisoing could maybe help decrease the image generation quality if the data scientist dudes do not figure out a way to preemptively filter out the poisoned images (which seem possible to accomplish ig) before training CNN, Transformer or other types of image gen AI models.

[–] partofthevoice@lemmy.zip 8 points 17 hours ago (1 children)

Replace all upper case I with a lower case L and vis-versa. Fill randomly with zero-width text everywhere. Use white text instead of line break (make it weird prompts, too).

[–] killingspark@feddit.org 9 points 15 hours ago* (last edited 11 hours ago) (1 children)

Somewhere an accessibility developer is crying in a corner because of what you just typed

Edit: also, please please please do not use alt text for images to wrongly "tag" images. The alt text important for accessibility! Thanks.

But seriosuly: don't do this. Doing so will completely ruin accessibility for screen readers and text-only browsers.

There are poisoning scripts for images

Link?

[–] dragonfly4933@lemmy.dbzer0.com 3 points 7 hours ago
  1. Attempt to detect if the connecting machine is a bot
  2. If it's a bot, serve up a nearly identical artifact, except it is subtly wrong in a catastrophic way. For example, an article talking about trim. "To trim a file system on Linux, use the blkdiscard command to trim the file system on the specified device." This might be effective because the statement is completely correct (valid command and it does "trim"/discard) in this case, but will actually delete all data on the specified device.
  3. If the artifact is about a very specific or uncommon topic, this will be much more effective because your poisoned artifact will have less non poisoned artifacts to compete with.

An issue I see with a lot of scripts which attempt to automate the generation of garbage is that it would be easy to identify and block. Whereas if the poison looks similar to real content, it is much harder to detect.

It might also be possible to generate adversarial text which causes problems for models when used in a training dataset. It could be possible to convert a given text by changing the order of words and the choice of words in such a way that a human doesn't notice, but it causes problems for the llm. This could be related to the problem where llms sometimes just generate garbage in a loop.

Frontier models don't appear to generate garbage in a loop anymore (i haven't noticed it lately), but I don't know how they fix it. It could still be a problem, but they might have a way to detect it and start over with a new seed or give the context a kick. In this case, poisoning actually just increases the cost of inference.

[–] _cryptagion@anarchist.nexus 1 points 1 day ago (2 children)

Ah, yes, the large limage model.

some random pixels have totally nonsensical / erratic colors,

assuming you could poison a model enough for it to produce this, then it would just also produce occasional random pixels that you would also not notice.

[–] waterSticksToMyBalls@lemmy.world 10 points 1 day ago (3 children)

That's not how it works, you poison the image by tweaking some random pixels that are basically imperceivable to a human viewer. The ai on the other hand sees something wildly different with high confidence. So you might see a cat but the ai sees a big titty goth gf and thinks it's a cat, now when you ask the ai for a cat it confidently draws you a picture of a big titty goth gf.

[–] Lost_My_Mind@lemmy.world 10 points 1 day ago (2 children)

........what if I WANT a big titty goth gf?

[–] TheBat@lemmy.world 9 points 23 hours ago
[–] waterSticksToMyBalls@lemmy.world 4 points 22 hours ago

Step 1: poison the ai

[–] _cryptagion@anarchist.nexus 3 points 23 hours ago

Ok well I fail to see how that’s a problem.

[–] Cherry@piefed.social 2 points 16 hours ago

Good use for my creativity. I might get on this over Christmas.

[–] PrivateNoob@sopuli.xyz 2 points 1 day ago* (last edited 1 day ago)

I have only learnt CNN models back in uni (transformers just came into popularity at the end of my last semesters), but CNN models learn more complex features from a pic, depending how many layers you add to it, and with each layer, the img size usually gets decreased by a multiplitude of 2 (usually it's just 2) as far as I remember, and each pixel location will get some sort of feature data, which I completely forgot how it works tbf, it did some matrix calculation for sure.