1319
submitted 6 months ago by boem@lemmy.world to c/technology@lemmy.world
you are viewing a single comment's thread
view the rest of the comments
[-] Brickardo@feddit.nl 1 points 6 months ago

Let's agree to disagree then. An LLM has no notion of semantics, it's just outputting the most likely word to follow up to what it's already written and the user's input.

On the contrary, expert systems from back in the 90s for, say, predicting the atomic structure of an element, work like a human brain on steroids. It features an arbitrary large search tree that the software knows how to iterarively prune according to a well known set of chemical rules. We do the same when analyzing a set of options.

Debugging "current" AI models, on the other hand, is impossible because all we're doing is prescripting a composition of functions and forcing it to minimize a loss function. That's all we're doing. How can you currently tell that a certain model is going to work? Unless the mathematical theory ever catches up with the technology, we'll never know until we execute the code.

this post was submitted on 03 Jun 2024
1319 points (100.0% liked)

Technology

59974 readers
1935 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 2 years ago
MODERATORS