this post was submitted on 23 Dec 2023
171 points (86.4% liked)

Technology

58143 readers
5215 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] huginn@feddit.it 166 points 9 months ago (5 children)

Friendly reminder that your predictive text, while very compelling, is not alive.

It's not a mind.

[–] Poggervania@kbin.social 75 points 9 months ago (2 children)

Cyberpunk 2077 sorta explores this a bit.

There’s a vending machine that has a personality and talks to people walking by it. The quest chain basically has you and the vending machine chatting a bit and even giving the vending machine some advice on a person he has a crush on. You eventually become friends with this vending machine.

When it seems like it’s becoming more apparent it’s an AI and is developing sentience, it turns out the vending machine just has a really well-coded socializing program. He even admits as much when he’s about to be deactivated.

So, to reiterate what you said: predictive text and LLMs are not alive nor a mind.

[–] dlpkl@lemmy.world 44 points 9 months ago* (last edited 9 months ago)

I don't care, Brandon was real to me okay 😭

[–] billwashere@lemmy.world 20 points 9 months ago

Which is why the Turing Test needs to be updated. These text models are getting really good at fooling people.

[–] CrayonRosary@lemmy.world 17 points 9 months ago (2 children)

Prove to me you have a mind and I'll accept what you're saying.

[–] LWD@lemm.ee 1 points 9 months ago* (last edited 8 months ago)
[–] huginn@feddit.it -5 points 9 months ago

Well there are 2 options:

Either I'm a real mind separate and independent of you or I'm a figment of your imagination.

At which point you have to ask yourself: why are you so convinced you're an unlovable and insufferable twat?

[–] JayDee@lemmy.ml 7 points 8 months ago (1 children)

I don't think most people will care, so long as their NPC interaction ends up compelling. We've been reading stories about people who don't exist for centuries, and that's stopped no one from sympathizing with them - and now there's a chance you could have an open conversation with them.

Like, I think alot of us assume that we care about the authors who write the character dialogs but I think most people actually choose not to know who is behind their favorite NPCs to preserve some sense that the NPC personality isn't manufactured.

Combine that with everyone becoming steadily more lonely over the years, and I think AI-generated NPC interactions are going to take escapism to another level.

[–] PsychedSy@sh.itjust.works 2 points 8 months ago (1 children)

Poem poem poem poem then the NPC start quoting Mein Kampf and killing all the cat wizards.

[–] JayDee@lemmy.ml 1 points 8 months ago

Lol, yeah. If generative AI text stays as shitty as it is now, then this whole discussion moot. Whether that will be the case has yet to be seen. What is an indisputable fact, though, is that right now is the worst that generative AI will ever be again. It's only able to improve from here.

[–] Bluehat@lemmynsfw.com 1 points 9 months ago* (last edited 9 months ago) (1 children)
[–] Bernie_Sandals@lemmy.world 0 points 8 months ago

If you cut out a tiny bit of someone's brain and then hooked it up to a cpu, would it be a mind? No, of course not, lol. Even if we got Biocomputers to work, we still wouldn't have any synthetic hardware even close to being strong or fast enough to actually create or even simulate a brain.

[–] MxM111@kbin.social 0 points 9 months ago (3 children)

While it is not alive, whether it is a mind is not a clear cut. It can be called kind of a mind, a mind different of that of human.

[–] match@pawb.social 5 points 9 months ago

What can't be a kind of mind to you?

[–] Corgana@startrek.website 1 points 8 months ago

Sorry you're getting downvoted, you're correct. It's not implausible to assume that generative AI systems may have some kind of umwelt, but it is highly implausible to expect that it would be anything resembling that of a human (or animal). I think people are getting hung up on it because they're assuming a lack of understanding language implies a lack of any concious experience. Humans do lots of things without understanding how they might be understood by others.

To be clear, I don't think these systems have experience, but it's impossible to rule out until an actual robust theory of mind comes around.

[–] huginn@feddit.it 0 points 9 months ago (3 children)

Unless you want to call your predictive text on your keyboard a mind you really can't call an LLM a mind. It is nothing more than a linear progression from that. Mathematically proven to not show any form of emergent behavior.

[–] kogasa@programming.dev 4 points 8 months ago (1 children)

No such thing has been "mathematically proven." The emergent behavior of ML models is their notable characteristic. The whole point is that their ability to do anything is emergent behavior.

[–] huginn@feddit.it 0 points 8 months ago* (last edited 8 months ago) (1 children)

Here's a white paper explicitly proving:

  1. No emergent properties (illusory due to bad measures)
  2. Predictable linear progress with model size

https://arxiv.org/abs/2304.15004

The field changes fast, I understand it is hard to keep up

[–] kogasa@programming.dev 2 points 8 months ago (1 children)

Sure, if you define "emergent abilities" just so. It's obvious from context that this is not what I described.

[–] huginn@feddit.it -1 points 8 months ago (1 children)

Their paper uses industry standard definitions

[–] kogasa@programming.dev 2 points 8 months ago

Their paper uses terminology that makes sense in context. It's not a definition of "emergent behavior."

[–] MxM111@kbin.social 3 points 9 months ago* (last edited 9 months ago) (1 children)

I do not think that it is “linear” progression. ANN by definition is nonlinear. Neither I think anything is “mathematically proven”. If I am wrong, please provide a link.

[–] huginn@feddit.it 1 points 8 months ago (1 children)

Sure thing: here's a white paper explicitly proving:

  1. No emergent properties (illusory due to bad measures)
  2. Predictable linear progress with model size

https://arxiv.org/abs/2304.15004

[–] MxM111@kbin.social 2 points 8 months ago* (last edited 8 months ago)

Thank you. This paper though does not state that there are no emergent abilities. It only states that one can introduce a metric with respect to which the emergent ability behaves smoothly and not threshold-like. While interesting, it only suggests that things like intelligence are smooth functions, but so what? Some other metrics show exponential or threshold dependence and whether the metric is right depends only how one will use it. And there is no law that emerging properties have to be threshold like. Quite the opposite - nearly all examples in physics that I know, the emergence appears gradually.

[–] General_Effort@lemmy.world 1 points 8 months ago (1 children)

It is obvious that you do not know what either "mathematical proof" or "emergence" mean. Unfortunately, you are misrepresenting the facts.

I don't mean to criticize your religious (or philosophical) convictions. There is a reason people mostly try to keep faith and science separate.

[–] huginn@feddit.it 0 points 8 months ago (1 children)

Here's a white paper explicitly proving:

No emergent properties (illusory due to bad measures)

Predictable linear progress with model size

https://arxiv.org/abs/2304.15004

The field changes fast, I understand it is hard to keep up

[–] General_Effort@lemmy.world 0 points 8 months ago (1 children)

As I said, you do not understand what these 2 terms mean. As such, you are incapable of understanding that paper.

Perhaps your native language is Italian, so here are links to the .it Wikipedia.

https://it.wikipedia.org/wiki/Comportamento_emergente

https://it.wikipedia.org/wiki/Dimostrazione_matematica

[–] huginn@feddit.it 1 points 8 months ago (1 children)
  1. Emergence is the whole being greater than the sum of its parts. That's the original meaning of emergent properties, which is laid out in the first paragraph of the article. It's the scholarly usage as well, and what the claims of observed emergence are using as the base of their claim.

  2. The article very explicitly demonstrated that only about 10% of any of the measures for LLMs displayed any emergence and that illusory emergence was the result of overly rigid metrics. Swapping to edit distance as an approximately close metric causes the sharp spikes to disappear for obvious reasons: no longer having a sharp yes/no allows for linear progression to reappear. It was always there, merely masked by flawed statistics.

If you can't be bothered to read here's a very easy to understand video by one of the authors: https://www.youtube.com/watch?v=ypKwNrmuuPM

[–] General_Effort@lemmy.world 0 points 8 months ago

Good. Now do you understand how you have misrepresented the paper?