this post was submitted on 25 Nov 2025
668 points (98.8% liked)

Technology

77072 readers
3010 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related news or articles.
  3. Be excellent to each other!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, this includes using AI responses and summaries. To ask if your bot can be added please contact a mod.
  9. Check for duplicates before posting, duplicates may be removed
  10. Accounts 7 days and younger will have their posts automatically removed.

Approved Bots


founded 2 years ago
MODERATORS
 

Generative “AI” data centers are gobbling up trillions of dollars in capital, not to mention heating up the planet like a microwave. As a result there’s a capacity crunch on memory production, shooting the prices for RAM sky high, over 100 percent in the last few months alone. Multiple stores are tired of adjusting the prices day to day, and won’t even display them. You find out how much it costs at checkout.

you are viewing a single comment's thread
view the rest of the comments
[–] drosophila@lemmy.blahaj.zone 10 points 1 day ago* (last edited 1 day ago)

Hallucinations are an intrinsic part of how LLMs work. OpenAI, literally the people with the most to lose if LLMs aren't useful, has admitted that hallucinations are a mathematical inevitability, not something that can be engineered around. On top of that, its been shown that for things like mathematical proof finding switching to more sophisticated models doesn't make them more accurate, it just makes their arguments more convincing.

Now, you might say "oh but you can have a human in the loop to check the AIs work", but for programming tasks its already been found that using LLMs makes programmers less productive. If a human needs to go over everything an AI generates, and reason about it anyway, that's not really saving time or effort. Now consider that as you make the LLM more complex, having it generate longer and more complicated blocks of text, its errors also become harder to detect. Is that not just shuffling around the necessary human brainpower for a task instead of reducing it?

So, in what field is this sort of thing useful? At one point I was hopeful that LLMs could be used in text summarization, but if I have to read the original text anyway to make sure that I haven't been fed some highly convincing falsehood then what is the point?

Currently I'm of the opinion that we might be able to use specialized LLMs as a heuristic to narrow the search tree for things like SAT solvers and answer set generators, but I don't have much optimism for other use cases.