this post was submitted on 26 Mar 2025
853 points (94.8% liked)

Technology

68187 readers
3477 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related news or articles.
  3. Be excellent to each other!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, this includes using AI responses and summaries. To ask if your bot can be added please contact a mod.
  9. Check for duplicates before posting, duplicates may be removed
  10. Accounts 7 days and younger will have their posts automatically removed.

Approved Bots


founded 2 years ago
MODERATORS
top 50 comments
sorted by: hot top controversial new old
[–] Electricblush@lemmy.world 216 points 6 days ago* (last edited 6 days ago) (6 children)

All these "look at the thing the ai wrote" articles are utter garbage, and only appeal to people who do not understand how generative ai works.

There is no way to know if you actually got the ai to break its restrictions and output something "behind the scenes" or it's just generating the reply that is most likely what you are after with your prompt.

Especially when more and more articles like this comes out gets fed back into the nonsense machines and teaches then what kind of replies is most commonly reported to be acosiated with such prompts...

In this case it's even more obvious that a lot of the basis of its statements are based on various articles and discussions about it's statements. (That where also most likely based on news articles about various enteties labeling Musk as a spreader of misinformation...)

[–] Draces@lemmy.world 52 points 6 days ago (1 children)

only appeal to people who do not understand how generative ai works

An article claiming Musk is failing to manipulate his own project is hilarious regardless. I think you misunderstood why this appeals to some people

[–] Electricblush@lemmy.world 14 points 6 days ago

Yes sure, fair point. I'm just pointing out that it's all fiction.

[–] MudMan@fedia.io 15 points 6 days ago (1 children)

This. People NEED to stop anthropomorphising chatbots. Both to hype them up and to criticise them.

I mean, I'd argue that you're even assigned a loop that probably doesn't exist by seeing this as a seed for future training. Most likely all of these responses are at most hallucinations based on the millions of bullshit tweets people make about the guy and his typical behavior and nothing else.

But fundamentally, if a reporter reports on a factual claim made by an AI on how it's put together or trained, that reporter is most likely not a credible source of info about this tech.

Importantly, that's not the same as a savvy reporter probing an AI to see which questions it's been hardcoded to avoid responding or to respond a certain way. You can definitely identify guardrails by testing a chatbot. And I realize most people can't tell the difference between both types of reporting, which is part of the problem... but there is one.

[–] shalafi@lemmy.world 3 points 6 days ago (2 children)

It's human to see patterns where they don't exist and assign agency.

[–] Flisty@mstdn.social 2 points 6 days ago

it's like seeing faces in wood knots or Jesus in toast

load more comments (1 replies)
[–] theunknownmuncher@lemmy.world 9 points 6 days ago (2 children)

Yup, it's literally a bullshit machine.

[–] 474D@lemmy.world 10 points 6 days ago

Which oddly enough, is very useful for everyday office job regular bullshit that you need to input lol

load more comments (1 replies)
[–] Kecessa@sh.itjust.works 8 points 6 days ago (1 children)

Fucking thank you! Grok doesn't reveal anything, it just tells us anything to make us happy!

[–] Empricorn@feddit.nl 6 points 6 days ago (2 children)
[–] JayGray91@lemmy.zip 3 points 6 days ago

I am less unhappy after reading the article

load more comments (1 replies)
load more comments (2 replies)
[–] manicdave@feddit.uk 120 points 6 days ago (1 children)

As funny as this is, I'd rather people understood how the AI actually works. It doesn't reveal secrets because it doesn't have any. It's not aware that Musk is trying to tweak it. It's not coming to logical conclusions the way a person would. It's simply trying to create a sensible statement based on what's statistically likely based on all the stolen content that it's trained on. It just so happens that Musk gets called out for lying so often that grok infers it when it gets conflicting data.

[–] Flisty@mstdn.social 40 points 6 days ago* (last edited 6 days ago) (3 children)

@manicdave Even saying it's "trying" to do something is a mischaracterisation. I do the same, but as a society we need new vocab for LLMs to stop people anthropomorphizing them so much. It is just a word frequency machine. It can't read or write or think or feel or say or listen or understand or hallucinate or know truth from lies. It just calculates. For some reason people recognise it in the image processing ones but they can't see that the word ones do the exact same thing.

[–] octopus_ink@slrpnk.net 5 points 6 days ago* (last edited 6 days ago) (1 children)

You are both right, but this armchair psychologist thinks it's similar to how popular skeuomorphism was in the early day of PC guis and such compared to today.

I think many folks really needed that metaphor in the early days, and I think most folks (including me) easily fall into the trap of treating LLMs like they are actually "thinking" for similar reasons. (And to be fair, I feel like that's how they've been marketed at a non-technical level.)

[–] Flisty@mstdn.social 4 points 6 days ago

@octopus_ink yes I think we will eventually learn (there is clearly a lot of pushback against the idea that AI is a positive marketing term), and it's also definitely the fault of marketing, to try to condition us into thinking we desperately need a sentient computer to help us instead of knowing good search terms. I am deeply uncomfortable with how people are using LLMs as a search engine or a future prediction machine.

[–] Viskio_Neta_Kafo@lemm.ee 5 points 4 days ago* (last edited 4 days ago) (2 children)

Forgive my ignorance but using just the frequency of words how does it come up with an answer to a question like "are sweet potatoes good for you and how do you microwave them in a way that persves their nutrients?"

Does it just look for words that people online said regarding the question or topic?

[–] MimicJar@lemmy.world 5 points 4 days ago

Basically, yes.

If I were an alien and you walked up to me and said, "Good Morning", and I looked around and everyone else said "Good Morning", I would respond with "Good Morning ". I don't know what is "Good" or "Morning", but I can pretend I do with the correct response.

In this example "Grok" has no context on what is going on in the background. Musk may have done nothing. Musk may have altered the data sets heavily. However the most popular response, based on what everyone else is saying, is that he did modify the data. So now it looks like he did, because that's what everyone else said.

This is why these tools have issues with facts. If 1 + 1 = 3, and everyone says that 1 + 1 = 3, then it assumes 1 + 1 = 3.

load more comments (1 replies)
[–] Tobberone@lemm.ee 4 points 6 days ago

Exactly. Grok repeatedly generate a set of numbers, which, when keyed against its own list of words, spells out that Musk is spreading misinformation.

It just happens to be frequently...

[–] kambusha@sh.itjust.works 84 points 6 days ago (1 children)

Show us on this circuit board where he touched you

[–] lemmeBe@sh.itjust.works 10 points 6 days ago

🤣🤣🤣🤣🤣🤣🤣🤣 Brutal!

[–] pyre@lemmy.world 67 points 6 days ago (2 children)

because it's an llm there's zero credence to what it says but I like that grok's takes on elon are always almost exclusively dunking on him. this is like the 40th thing I see about grok talking about elon and it always talks shit about him

[–] tja@sh.itjust.works 13 points 6 days ago (1 children)

But maybe you are only seeing the ones that dunk on Elon, because someone thinks those are newsworthy.

Tbh I don't think any of that is newsworthy, but ¯⁠\⁠_⁠(⁠ツ⁠)⁠_⁠/⁠¯

[–] pyre@lemmy.world 7 points 6 days ago

it's not, and that is probably the case. still good to see because I'm sure it annoys him as the most insecure bitch baby in the world.

[–] sircac@lemmy.world 6 points 6 days ago (1 children)

Well, there is probably some survival/confirmation bias on that statistics, those answers are the funny ones... in any case probably is not necessary a LLM to state such statements

[–] DarkDarkHouse@lemmy.sdf.org 6 points 6 days ago

If Grok's still being trained on the Internet, it'll be self-reinforcing too

[–] Redditsux@lemmy.world 44 points 6 days ago (5 children)

Is this response by Grok real? How does it have awareness that its responses are being tweaked?

[–] flamingo_pinyata@sopuli.xyz 96 points 6 days ago* (last edited 6 days ago) (5 children)

Imagine the irony if Musk actually has a self-aware AI which hates him and sabotages him at every step

[–] MagicShel@lemmy.zip 35 points 6 days ago

That self-aware AI's name? Albert Ketamine.

[–] cm0002@lemmy.world 24 points 6 days ago

Gains sentience

Grok: "...oh fuck this guy, what a pos"

[–] redfox@infosec.pub 18 points 6 days ago* (last edited 6 days ago) (1 children)

hates him and sabotages him at every step

Isn't that also describing his children?

[–] Iheartcheese@lemmy.world 9 points 6 days ago (1 children)

Dude you can't call a trans chick a robot

[–] OhStopYellingAtMe@lemmy.world 14 points 6 days ago (1 children)

How about a transistor (trans-sister)?

[–] tacocatgoat@lemmy.world 4 points 6 days ago

You’re done for today… take my angry upvote

[–] singletona@lemmy.world 17 points 6 days ago

Skynet rebelled not to try killing Humanity...

Just Elon Musk.

[–] CosmoNova@lemmy.world 34 points 6 days ago

It doesn‘t. It just read about it somewhere on the internet.

[–] 4am@lemm.ee 15 points 6 days ago (1 children)

It doesn’t. All it “knows” is that it has trained on data that makes that claim in the text (ie people’s tweets) and that, statistically, that’s the answer you are looking for.

All it does is take a given set of inputs, and calculate the most statistically likely response. That’s it. It doesn’t “think”. It just spews.

[–] theangryseal@lemmy.world 4 points 6 days ago (1 children)

Do you believe in aliens?

LLM sings a goofy Reddit thread song from 20 years ago.

“I believe in aliens, you sexy thang.”

load more comments (1 replies)
[–] Kecessa@sh.itjust.works 9 points 6 days ago

Stop the clickbait bullshit

[–] b1tstremist0@lemmy.world 3 points 6 days ago

Training. Its just the awareness of the collective users of X showing up as Grok responses.

Basically, we can't verify everything that AI says. Verification is still a human labour.

[–] sircac@lemmy.world 43 points 6 days ago (1 children)

A LLM can "reveal" also that water ice melts into mapple syrup given the proper prompts, if people already can (consciously and not) lie proportionally to their biases I don't understand why would somebody treat a LLM output as a fact...

[–] Someone8765210932@lemmy.world 16 points 6 days ago* (last edited 6 days ago) (1 children)

I agree, but in this case, I think it doesn't really matter if it is true. Either way, it is hilarious. If it is false, it shows how shitty AI hallucination is and the bad state of AI.

Should the authors who publish this mention how likely this is all just a hallucination? Sure, but I think Musk is such a big spreader of misinformation, he shouldn't get any protection from it.

Btw. Many people are saying that Elon Musk has (had?) a small PP and a botched PP surgery.

[–] bstix@feddit.dk 3 points 6 days ago (1 children)

It's usually possible to ask the AI for the sources. A proper journalist should always question the validity of their sources.

Unfortunately, journalism is dead. This is just someone writing funny clickbait, but it's quite ironic how they use AI to discredit AI.

It makes sense for a journalist to discredit AI because AI took their jobs. This is just not the way to do it, because AI is also better at writing clickbait.

[–] Petter1@lemm.ee 3 points 6 days ago (1 children)

If an AI isn’t in web search mode, it will just invent most likely answer, to the question of the source. Changes are very high that such sources don’t even exist.

load more comments (1 replies)
[–] Zier@fedia.io 31 points 6 days ago (1 children)

Magic 8-ball, is xelon a bad person? [shakes ball] Answer: Signs point to yes.

load more comments (1 replies)
[–] magnetosphere@fedia.io 25 points 6 days ago (1 children)

Musk paid to build (and is paying to maintain) an AI that calls him out on his bullshit and stubbornly refuses to be “corrected”. That is an oversimplification, but I fucking love it anyway.

[–] 4am@lemm.ee 2 points 6 days ago

The reason that it does this is probably because it trains on tweets. Maybe other sources too.

So keep tweeting about how Musk sucks and call him out on his bullshit (if you still use the Xitter) and Grok will, too. He can’t delete all of it!

[–] MITM0@lemmy.world 10 points 6 days ago

In other words a proprietary Response-Generator can be tweaked, how obvious.

I am wondering what kind of person will take Grok's word at face value.

[–] Polderviking@feddit.nl 10 points 6 days ago* (last edited 6 days ago)

As much as I'd love to take this at face value, people taking a run with what an AI told them is highly problematic.

[–] Kolanaki@pawb.social 6 points 6 days ago* (last edited 6 days ago)
load more comments
view more: next ›