this post was submitted on 30 Oct 2025
916 points (99.4% liked)

Technology

76512 readers
2272 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related news or articles.
  3. Be excellent to each other!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, this includes using AI responses and summaries. To ask if your bot can be added please contact a mod.
  9. Check for duplicates before posting, duplicates may be removed
  10. Accounts 7 days and younger will have their posts automatically removed.

Approved Bots


founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[โ€“] definitemaybe@lemmy.ca 2 points 23 hours ago* (last edited 20 hours ago) (1 children)

Re: your last paragraph:

I think the future is likely going to be more task-specific, targeted models. I don't have the research handy, but small, targeted LLMs can outperform massive LLMs at a tiny fraction of the compute costs to both train and run the model, and can be run on much more modest hardware to boot.

Like, an LLM that is targeted only at:

  • teaching writing and reading skills
  • teaching English writing to English Language Learners
  • writing business emails and documents
  • writing/editing only resumes and cover letters
  • summarizing text
  • summarizing fiction texts
  • writing & analyzing poetry
  • analyzing poetry only (not even writing poetry)
  • a counselor
  • an ADHD counselor
  • a depression counselor

The more specific the model, the smaller the LLM can be that can do the targeted task (s) "well".

Yeah I agree. Small models is the way. You can also use LoRa/QLoRa adapters to "fine tune" the same big model for specific tasks and swap the use case in realtime. This is what apple do with apple intelligence. You can outperform a big general LLM with an SLM if you have a nice specific use case and some data (which you can synthesise in come cases)