this post was submitted on 15 Apr 2026
-48 points (23.3% liked)

Technology

83966 readers
4517 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related news or articles.
  3. Be excellent to each other!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, this includes using AI responses and summaries. To ask if your bot can be added please contact a mod.
  9. Check for duplicates before posting, duplicates may be removed
  10. Accounts 7 days and younger will have their posts automatically removed.

Approved Bots


founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] brynden_rivers_esq@lemmy.ca 5 points 4 days ago (2 children)

I'll have to take your word for it! "figuring out" sounds like a higher-order process than a large language model is capable of to me, but if what they do is as good, then great.

I think I'm just skeptical because of how horrendously bad LLM output is in my field of expertise (despite looking fine to a lay person), so I immediately analogize that to other areas. The output of law and coding are both really about language, and the process of creating that output on the part of a lawyer or coder are really about language, so I can see how one might think LLMs would be able to recreate what lawyers and coders do. But boy it doesn't strike me as remotely plausible that LLMs will ever get there, at least for law. I have no doubt some yet-unimagined technology could get us there, but "next word prediction" just isn't gonna be it.

[–] phutatorius@lemmy.zip 2 points 4 days ago (1 children)

The more specialized and less public the knowledge is that's needed to train an LLM, the worse its output will be. In addition, explainability is an absolute necessity where safety is a concern, but LLMs are not good at explaining how they got their results (because the results are derived from a statistical process, not logical steps originating from first principles). I suspect that explainability and verifiability are also essential in law.

[–] brynden_rivers_esq@lemmy.ca 1 points 4 days ago

They are, for sure. I mean in some sense, the explain-ability is why it's correct...you might need to explain why it's correct to a judge one day!