this post was submitted on 22 Jul 2024
190 points (96.1% liked)

Technology

59679 readers
4751 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] Diplomjodler3@lemmy.world -1 points 4 months ago* (last edited 4 months ago) (2 children)

People always assume that the current state of generative AI is the end point. Five years ago nobody would have believed what we have today. In five years it'll all be a different story again.

[–] Eccitaze@yiffit.net 7 points 4 months ago* (last edited 4 months ago) (1 children)

People always assume that generative AI (and technology in general) will continue improving at the same pace it always has been. They always assume that there are no limits in the number of parameters, that there's always more useful data to train it on, and that things like physical limits in electricity infrastructure, compute resources, etc., don't exist. In five years generative AI will have roughly the same capability it has today, barring massive breakthroughs that result in a wholesale pivot away from LLMs. (More likely, in five years it'll be regarded similarly to cryptocurrency is today, because once the hype dies down and the VC money runs out the AI companies will have to jack prices to a level where it's economically unviable to use in most commercial environments.)

[–] Zron@lemmy.world 3 points 4 months ago

To add to this, we’re going to run into the problem of garbage in, garbage out.

LLMs are trained on text from the internet.

Currently, a massive amount of text on the internet is coming from LLMs.

This creates a cycle of models getting trained on data sets that increasingly contain large sets of data generated by older models.

The most likely outlook is that LLMs will get worse as the years go by, not better.

[–] btaf45@lemmy.world 6 points 4 months ago

In five years it’ll all be a different story again.

You don't know that. Maybe it will take 124 years to make the next major breakthru and until then all that will happen is people will tinker around and find that improving one thing makes another thing worse.