this post was submitted on 12 Jul 2024
305 points (100.0% liked)

Technology

37801 readers
200 users here now

A nice place to discuss rumors, happenings, innovations, and challenges in the technology sphere. We also welcome discussions on the intersections of technology and society. If it’s technological news or discussion of technology, it probably belongs here.

Remember the overriding ethos on Beehaw: Be(e) Nice. Each user you encounter here is a person, and should be treated with kindness (even if they’re wrong, or use a Linux distro you don’t like). Personal attacks will not be tolerated.

Subcommunities on Beehaw:


This community's icon was made by Aaron Schneider, under the CC-BY-NC-SA 4.0 license.

founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] jarfil@beehaw.org 3 points 5 months ago* (last edited 5 months ago) (1 children)

It's not a statistical method anymore. One of the breakthroughs of large model neural networks, has been that during training an emergent process, assigns neurons to both relatively high level and specific traits, which at the same time "cluster up" with other neurons assigned to related traits. Adding just a bit of randomness ("temperature") allows the AI to jump from activating one trait to a close one, but not to one too far away. Confidence becomes a measure of how close is the output, to a consistent set of traits trained into the network. Interestingly, a temperature of 0 gives a confidence of 100%... but produces gibberish.

If its data contains a commonly held belief, that is incorrect

This is where things start to get weird. An AI system based on an LLM, can iterate over its own answers looking for the optimal one (Q*), and even detect inconsistencies in them. What it does after that, depends on whoever programmed it:

  • Maybe it casts any doubt aside, and outputs the first answer anyway (original ChatGPT did that, didn't even bother self-checking too much)
  • Or it could ask an authoritative source (ChatGPT plugins work like that)
  • Or it could search the web for additional info (Copilot and Gemini do that)
  • Or it could alert the user to both the low confidence and the inconsistencies (...but people want omniscient AIs, not "err... I'm not sure, Dave" AIs)
  • ...or, sometime in the future (or present?) they could re-train themselves, maybe via generating a LoRa, that would bring in corrected biases, or even additional concepts.

Over time, I think different AI systems will evolve to target accuracy, consistency, creativity, etc. Current systems are kind of rudimentary compared to what's yet to come, and too many are used in very rudimentary ways by anyone who can slap an "AI" label and sell them.

[–] coffeetest@beehaw.org 1 points 5 months ago (1 children)

That is pretty interesting and thanks for posting it. I hear the words and its intriguing but to be honest, I don't really understand it. I'd have to give it some thought and read more about it. Do you have a place you suggest going to learn more?

I use chatgpt-4o currently for learning python and helping with grammar. I find it does great with grammar but even with relatively simple python questions it can produce some "creative" answers. Like its in the ball park but its not perfect and for a learner, that's learning the hard way. To be fair I don't use the assistant/code interpreter, which I have no idea about but based on its name I assume it might be better. So that's what I based my somewhat skeptical opinion of ai on.

[–] jarfil@beehaw.org 2 points 5 months ago

Check out this one for a general overview:

https://youtu.be/OFS90-FX6pg

You may want to also check an intro to neural networks, and Q* is a somewhat new concept. Other than that... "the internet". There are plenty of places with info, not sure if there is a more centralized and structured one.

Learning to code with just ChatGPT is not the best idea. You need to join three areas:

  • general principles (data structures, algorithms, etc)
  • language rules (best described in a language reference)
  • business logic (computer science, software engineering, development patterns, etc)

ChatGPT's programming answers, give you an intersection of all those, often with some quirks, with the nice but only benefit of explaining what it thinks it is doing. You still need to have some basic understanding of those in order to understand what ChatGPT is talking about, how to double-check it, and how to look for more info. It can be a great timesaver as a way to generate drafts, though.