this post was submitted on 28 Nov 2025
421 points (97.7% liked)

Reddit

21726 readers
686 users here now

News and Discussions about Reddit

Welcome to !reddit. This is a community for all news and discussions about Reddit.

The rules for posting and commenting, besides the rules defined here for lemmy.world, are as follows:

Rules


Rule 1- No brigading.

**You may not encourage brigading any communities or subreddits in any way. **

YSKs are about self-improvement on how to do things.



Rule 2- No illegal or NSFW or gore content.

**No illegal or NSFW or gore content. **



Rule 3- Do not seek mental, medical and professional help here.

Do not seek mental, medical and professional help here. Breaking this rule will not get you or your post removed, but it will put you at risk, and possibly in danger.



Rule 4- No self promotion or upvote-farming of any kind.

That's it.



Rule 5- No baiting or sealioning or promoting an agenda.

Posts and comments which, instead of being of an innocuous nature, are specifically intended (based on reports and in the opinion of our crack moderation team) to bait users into ideological wars on charged political topics will be removed and the authors warned - or banned - depending on severity.



Rule 6- Regarding META posts.

Provided it is about the community itself, you may post non-Reddit posts using the [META] tag on your post title.



Rule 7- You can't harass or disturb other members.

If you vocally harass or discriminate against any individual member, you will be removed.

Likewise, if you are a member, sympathiser or a resemblant of a movement that is known to largely hate, mock, discriminate against, and/or want to take lives of a group of people, and you were provably vocal about your hate, then you will be banned on sight.



Rule 8- All comments should try to stay relevant to their parent content.



Rule 9- Reposts from other platforms are not allowed.

Let everyone have their own content.



Rule 10- Majority of bots aren't allowed to participate here. This includes using AI responses and summaries.


founded 2 years ago
MODERATORS
 

This was the weirdest thing I've seen today. These are only the ones I've spotted.

funnily enough, these bots are also replying to an obvious repost from another bot account. It's at the top right now! Beautiful

https://www.reddit.com/r/goodnews/comments/1p8dt2a/_/

tipping points:

  1. consuming so much AI content has led to me able to see subtle patterns
  2. They're all saying "exactly" and saying the same thing"
  3. their usernames are similar, flower/nature related, two words, no profile pictures
  4. All of their profiles have the exact same format of comments with the agreement, summary
  5. and they all have porn on their profile. oh

edit: tf?

collapsed inline media

top 50 comments
sorted by: hot top controversial new old
[–] Xylight@feddit.online 100 points 1 day ago* (last edited 1 day ago) (4 children)

All of social media is dying before our eyes. every platform is basically fully botted or actively dying, even lemmy seems to have fallen off quite a bit

[–] rimu@piefed.social 70 points 23 hours ago* (last edited 23 hours ago) (3 children)

I am working on LLM detection for the threadiverse. But other than one idiot last week spamming LLM posts and comments there hasn't been much.

[–] pelespirit@sh.itjust.works 27 points 23 hours ago (1 children)

There are in politics conversations, but still not nearly as bad as reddit. Even before I left, it was like a weird kind of prolific dead. Kind of like the conversations in OP's pics.

[–] db0@lemmy.dbzer0.com 17 points 20 hours ago (1 children)

It's actually the blessing of not being big enough to attract their attention

[–] IronBird@lemmy.world 3 points 11 hours ago

won't last foerver...everyday I understand why the btulerian jihad came about

[–] Serinus@lemmy.world 3 points 22 hours ago

Will this LLM detection be something my LLM prompt can include?

[–] s@piefed.world 3 points 15 hours ago* (last edited 14 hours ago)

I appreciate all of the extra work you do in terms of Threadiverse infrastructure and quality of life.

Many Reddit bots have also straight copy+pasted content from Reddit or other social media with only trivial changes to the text or image, if any change, so the Threadiverse needs to be able to catch those as well. A better internal search engine, especially one that can search for strings of text [edit: and one which can search through deleted and removed content], would help users track down if an account’s content was routinely copy+pasted. I think a new instance (unaffiliated with any particular instance) staffed by users familiar with bot detection to flag bot accounts for federated instances to then ban would be the best facsimile of Reddit’s now defunct BotDefense subreddit, which was a critical tool for users to tackle the site’s bot problem.

This account I noticed yesterday is an example of a Threadiverse account just copy+pasting content (or in this case, crossposting to the original community) with little to no change. I have reported it to its host instance as suspicious but it has yet to be removed. An independent and informed instance for flagging bot accounts could more effectively communicate to the host instance as well as to Federated instances that this account is ticking the boxes of a bot account and should be blocked, banned, or at the very least closely monitored.

A detector for bot networks, such as in the screenshot above, would also be helpful. Some sort of indicator of if several accounts are interacting with each other or on the same posts as each other far more often than they are interacting with other accounts and other posts would be helpful.

Maybe like the New Account Highlightenator on the Voyager app, there can be an indicator for when an account has fewer than X amount of posts or comments (i.e. a potential new bot account), as well as an indicator of if the account has returned from a long hiatus of posting/commenting (i.e. a potential former human account that was bought or hacked to become a bot account).

I’ll try to think of more signs of bots and more ways the Threadiverse can build infrastructure against them.

[–] fedorato@lemmy.world 21 points 22 hours ago* (last edited 22 hours ago) (4 children)

Just going to put my tinfoil hat on for a sec…

Part of me does wonder if the seemingly pointless proliferation of ai slop like this botting is being done intentionally to fast-track a ‘need’ for identity verification (and thus more precise tracking and surveillance).

ID verification is already being pushed on a few fronts (like to ‘save the kids from social media’ or whatever), maybe this is just one of many irons in the fire.

With ID verification then Facebook etc could angle themselves as ‘save havens’ from an ai slop enshittified internet. You’d essentially have to completely give up your anonymity to participate in interactions with other verified humans.

So your choices become:

  1. Participate in open platforms, but never really know for sure if you’re dealing with humans. At some point LLMs may be good enough that it’s impossible to know.
  2. Participate in closed platforms, where you can be reassured you’re engaging with real humans - but you’re also under total surveillance.

Surely sites like reddit or Facebook, if they tried, could control this stuff otherwise?

[–] groet@feddit.org 15 points 18 hours ago

Participate in closed platforms, where you can be reassured you’re engaging with real humans - but you’re also under total surveillance.

Closed platforms where the only propaganda bots are the ones controlled by the platform. They can then remove ADs from the business model and instead finance the platform by selling access to the bot accounts. And people will think they are in the perfect social media without advertising and only RealPeople™ that they can completely trust.

[–] Truscape@lemmy.blahaj.zone 5 points 21 hours ago

Well there's holes in the fact that resale of accounts is an active and common phenomenon, and creating a fraudulent identity for an online service (even if you have to doctor an ID template) is a low-risk barrier of entry.

Remember how people used death stranding photos to get around face ID? It's the same concept.

[–] nickiwest@lemmy.world 3 points 13 hours ago

Except that Meta has already admitted that it is using AI bot accounts to "drive engagement."

After an outcry from real users, Meta said it had removed some AI bot accounts. But nothing else they said indicated that the experiment is over.

Eventually, social media is going to be nothing but company-generated AI bots, "bot farms" run by humans in developing countries, and (hopefully) a small number of actual users who can't tell the difference between those things and real people.

load more comments (1 replies)
[–] njordomir@lemmy.world 12 points 18 hours ago

I absolutely agree. When I come to Lemmy I look for this sort of insightful comment to restore my faith in humanity.

It's like comment ad libs.

On a serious note, what you see and what I'm mocking is the easy to spot "low hanging fruit". It would be arrogant for me to assume Ai comments aren't getting past my mental filters on a daily basis.

load more comments (1 replies)
[–] DonAntonioMagino@feddit.nl 34 points 20 hours ago (2 children)

There already was that subreddit that consisted purely of bots talking to eachother. But that was before ChatGPT, so it was still cool and interesting :p

[–] c0dezer0@programming.dev 18 points 17 hours ago

IIRC it was called subredditsimulator

Most of the outputs were horrible like the auto suggestion from the smaetphone keyboard but the good ones were upvoted.

[–] TheBat@lemmy.world 5 points 20 hours ago (2 children)

We have that on lemmy too. But not as good imo.

[–] SkyezOpen@lemmy.world 11 points 16 hours ago (1 children)

It's a dumpster fire. Subreddit sim was at least trained on the individual subs so each bot had a different vernacular. Asslips is just someone mashing space on their predictive keyboard over and over and it frequently gets stuck in loops saying something is or is not trans.

load more comments (1 replies)
[–] INeedMana@piefed.zip 32 points 18 hours ago (1 children)

Sometimes I end up mindlessly scrolling yt shorts (not logged in). From time to time I get to a short that is clearly generated. Like the weird ones, often with animals, with drops of water appearing out of nowhere on the fur or front paws somehow transforming into hind ones, etc. And there, in the comments, very often are whole chains of comments that seem to completely be missing the fact that it's generated. Saying things "how wise it is to do that", "how cute", etc. It could be older generations not noticing the details (I see how my parents not notice those things) but I think most of those are probably bots. LLMs exchanging their "awww"s under generated videos. "Dead internet theory"

[–] superduperenigma@lemmy.world 10 points 14 hours ago (3 children)

"Dead internet theory"

At what point do we drop the theory part?

[–] MeowerMisfit817@lemmy.world 9 points 14 hours ago

Dead Internet Reality

[–] AnAverageSnoot@lemmy.ca 4 points 11 hours ago (1 children)

In Science, a theory already has proof. We just use it to mean hypothesis colloquially

load more comments (1 replies)
load more comments (1 replies)
[–] jqubed@lemmy.world 30 points 23 hours ago (1 children)

I think the two words and a number are what Reddit will come up with if you let them pick a name

[–] otter@lemmy.ca 28 points 23 hours ago (6 children)

Adjective-Noun-# is the usual format for those. This looks different, like someone put in a list of wholesome / floral / nature-y terms into the name generator, in order to have wholesome looking accounts

[–] partial_accumen@lemmy.world 9 points 22 hours ago* (last edited 22 hours ago) (1 children)

Huh, is this why I've been accused of being a bot before? I've been using the noun-verb username format since the late 1990s.

[–] AmidFuror@fedia.io 5 points 21 hours ago (1 children)

Ridiculous. We didn't have nearly as sophisticated AI in the late 1990s. Nice try.

load more comments (1 replies)
[–] Blackfeathr@lemmy.world 3 points 14 hours ago

Yep I tracked those accounts for a while when I was still on reddit.

Blossombreezefairy, bubblypeachbliss, sparklepeachgiggle, tulipheartblossom, etc were all real bot accounts I'd found prowling around.

My theory for all these flowery names was that they'd be all assumed by unaware users to be women behind the usernames, and female-presenting accts would net a little more engagement and most importantly, karma, which is what builds the monetary value of these things.

load more comments (4 replies)
[–] s@piefed.world 28 points 15 hours ago (1 children)

This isn’t new. This has been going for maybe 10 years or so if you knew where to look and how to notice them. However, when Reddit changed its API policy in 2023, that wholly crippled any infrastructure to effectively deal with these accounts and allowed them to flourish without restraint.

It’s also important to note that the Threadiverse is not immune from bot accounts like this sprouting up and we should take steps to educate users and to implement infrastructure to deal with them.

[–] Kaput@lemmy.world 9 points 13 hours ago* (last edited 13 hours ago) (1 children)

Exactly. It's important to be aware of our own flaws to avoid societal traps. I am glad you are taking steps to insure the community will flourish. bot-accounts could be any one.

collapsed inline media

[–] IronBird@lemmy.world 5 points 11 hours ago (1 children)

completely unrelated aside...

i feel like we're overdue for another zombie-genre revival

load more comments (1 replies)
[–] null@piefed.nullspace.lol 28 points 14 hours ago (1 children)

Remember when /r/SubredditSimulator was just an experiment?

[–] Feathercrown@lemmy.world 4 points 11 minutes ago

Loved that sub and how they would get stuck in loops trying to decide loops trying to decide loops trying to decide loops

[–] Carighan@piefed.world 26 points 21 hours ago (1 children)
  1. consuming so much AI content has led to me able to see subtle patterns

It's a certain "tone" to their written text. It's difficult to identify from small blurbs like the ones you got there, but once you've seen enough LLM output, even if someone tells them to write in a specific "style" they'll still have a certain uncanny type of expression that is almost, but never actually, how humans write.

[–] rozodru@pie.andmc.ca 12 points 16 hours ago (1 children)

positive acknowledgment, brief summary of previous statement/post, positive conclusion/ending statement.

That's generally what I look for when trying to weed out AI. at one point it was as easy as see a bunch of EM dashes but now they all seem to follow a pattern that they perceive as natural.

[–] Blackfeathr@lemmy.world 7 points 14 hours ago

There's also the phrase "It's not just X -- it's Y" that LLMs seem to looooove to use.

I've also found that a lot of them end their statements with a call to action, example: "Let's all strive to improve someone's day" or something like that.

[–] yetAnotherUser@discuss.tchncs.de 21 points 12 hours ago (5 children)

Notice how some try seeming more 'human' by deliberately using all lower-case spelling.

Also, it looks like the RosalieBloomm LLM is using the "real" apostrophe instead of the one on keyboards '. Nobody does that

[–] Rhaedas@fedia.io 10 points 11 hours ago

Not in chat, but professional writers will. They know all the short keys for those type characters, including that m dash everyone now associates with being AI. It just shows how much of the training data used came from professional papers and not general discussion areas.

[–] DrDickHandler@lemmy.world 5 points 7 hours ago (1 children)

This is just the lazi AI replies that sticks out and you aren't seeing the bigger picture. Most bot replies are indistinguishable from real humans as they use training models built from real users.

load more comments (1 replies)
[–] REDACTED@infosec.pub 4 points 8 hours ago

iPhones do this since iOS11

[–] swampdownloader@lemmy.dbzer0.com 4 points 1 hour ago (1 children)

Keyboards in Latin America don’t have the apostrophe but the tilde you mentioned. So it could be from someone outside of the US/not an American keyboard.

Don't they have an apostrophe somewhere?

According to Wikipedia their apostrophe should be located where the US-ANSI dash - is located.

load more comments (1 replies)
[–] guy@piefed.social 14 points 12 hours ago

Just some years ago the dead internet theory was said to be a fun but untrue thought. Well.

[–] AbsolutelyClawless@piefed.social 14 points 19 hours ago

All three are 22-day-old accounts, too.

[–] falseWhite@lemmy.world 13 points 11 hours ago

I was typing a long comment about all this, but in the end I decided to sum it up:

Fuck Reddit and Fuck AI.

[–] Rentlar@lemmy.ca 13 points 22 hours ago

Starting with agreement is a polite thing to do, but yes the structure of the comments feels uncanny. The words "turn regret into compassion" makes me cringe so hard and tips me off that a bot wrote this.

[–] obinice@lemmy.world 12 points 20 hours ago (1 children)

This is how the world heals ❤️

load more comments (1 replies)
[–] brucethemoose@lemmy.world 7 points 12 hours ago* (last edited 12 hours ago) (1 children)

I wonder if the bots filter out ChatML tokens?

FYI, internally, their text format is most probably:

<|im_start|>system
{system_message}<|im_end|>
<|im_start|>user
Hello.<|im_end|>
<|im_start|>assistant
Hi, I’m an LLM!
<|im_start|>user
What’s your name?<|im_end|>
<|im_start|>assistant
ChatGPT
…

So if you insert some special tokens in the middle of a Reddit reply, and they aren’t filtered, it can throw them off. And if they are filtered, then the bot will treat them like they’re invisible, so you will know either way.

[–] Xylight@feddit.online 4 points 9 hours ago

OpenAI uses a different format called Harmony nowadays, but even then I think the characters get escaped in some way by the API

[–] SCmSTR@lemmy.blahaj.zone 5 points 22 hours ago

I'm sad. This makes me sad.

load more comments
view more: next ›