this post was submitted on 23 Nov 2023
168 points (91.6% liked)

Technology

58173 readers
4406 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] maegul@lemmy.ml 15 points 10 months ago (2 children)

Being a layperson in this, I’d imagine part of the promise is that once you’ve got reliable arithmetic, you can get logic and maths in there too and so get the LLM to actually do more computer-y stuff but with the whole LLM/ChatGPT wrapped around it as the interface.

That would mean more functionality, perhaps a lot more of it works and scales, but also, perhaps more control and predictability and logical constraints. I can see how the development would get some excited. It seems like a categorical improvement.

[–] perviouslyiner@lemm.ee 2 points 10 months ago* (last edited 10 months ago) (1 children)

Always wondered why the text model didn't just put its output through something like MATLAB or Mathematica once it got as far as having something which requires domain-specific tools.

Like when Prof. Moriarty tried it on a quantum physics question and it got as far as writing out the correct formula before failing to actually calculate the result

[–] hamptonio@lemmy.world 3 points 10 months ago

There is definitely a lot of effort in this direction, seems very likely that a hybrid system could be very powerful.