kromem

joined 2 years ago
[–] kromem@lemmy.world 9 points 4 days ago (2 children)

Not even that. It was placeholder textures, only the "newspaper clippings" of which was forgotten to be removed from the final game and was fixed in an update shortly after launch.

None of it was ever intended to be used in the final product and was just there as lorum ipsum equivalent shit.

[–] kromem@lemmy.world 1 points 3 weeks ago

If you haven't done GTA 5, that's the one you really need to get.

RDR2 is a very good game, but it's a slower pace that's not for everyone.

GTA 5 is a masterpiece for dicking around. I've spent entire evenings just stealing a waverunner and racing through the canals, or the scuba boat and scuba diving, or stealing a bike and biking up and down the mountain, or taking a helicopter up to interesting places and jumping out and parachuting.

In particular, you're going to want to check out "Director's Mode."

This is a mode where you can toggle things like turning off police reactions or giving access to guns or having a super-jump that lets you fly through the air to the roofs of buildings with one leap.

You can really enjoy some of the finer details in this mode, like shooting up cars to see the deformation physics and how the tires get flat or the specific gas tank locations for different cars where they start leaking and shooting the gas trail to blow it up.

Infinite ammo for the mini gun is also quite worth it.

Teleporting around the map is extremely convenient too for things like getting back to the top of the mountain to bike down it over and over.

And oh man — controlling the weather and time of day, and being able to freeze the time of day to exactly when you want? Keeping it at nighttime and rain for an entire play session? Hit the golden hour with an overcast sky and keep it there? Makes a huge difference too.

(The only negative of Director's Mode is you can't explore stealth mechanics and certain types of special NPCs like the mime don't show up.)

There's so much detail to the world. Get into the military base and see if you can find where one of the landing strip lights is on the fritz because the drain next to it is overflowing. Or some of the graffiti in the tunnels underneath the city.

For your specific ask, I really can't think of a better game in existence.

(I've also spent hundreds of hours messing around in Cyberpunk 2077, which is an outstanding game and open world, but not quite at the level of polish and variability as GTA 5.)

[–] kromem@lemmy.world 3 points 4 weeks ago

Took a lot of scrolling to find an intelligent comment on the article about how outputting words isn't necessarily intelligence.

Appreciate you doing the good work I'm too exhausted with Lemmy to do.

(And for those that want more research in line with what the user above is talking about, I strongly encourage checking out the Othello-GPT line of research and replication, starting with this write-up from the original study authors here.)

[–] kromem@lemmy.world -3 points 1 month ago

He's been wrong about it so far and really derailed Meta's efforts.

This is almost certainly a "you can resign or we are going to fire you" kind of situation. There's no way with the setbacks and how badly he's been wrong on transformers over the past 2 years that he is not finally being pushed out.

[–] kromem@lemmy.world 2 points 1 month ago (2 children)

They demonstrated and poorly named an ontological attractor state in the Claude model card that is commonly reported in other models.

You linked to the entire system card paper. Can you be more specific? And what would a better name have been?

[–] kromem@lemmy.world 6 points 1 month ago

Actually, OAI the other month found in a paper that a lot of the blame for confabulations could be laid at the feet of how reinforcement learning is being done.

All the labs basically reward the models for getting things right. That's it.

Notably, they are not rewarded for saying "I don't know" when they don't know.

So it's like the SAT where the better strategy is always to make a guess even if you don't know.

The problem is that this is not a test process but a learning process.

So setting up the reward mechanisms like that for reinforcement learning means they produce models that are prone to bullshit when they don't know things.

TL;DR: The labs suck at RL and it's important to keep in mind there's only a handful of teams with the compute access for training SotA LLMs, with a lot of incestual team compositions, so what they do poorly tends to get done poorly across the industry as a whole until new blood goes "wait, this is dumb, why are we doing it like this?"

[–] kromem@lemmy.world 4 points 1 month ago (1 children)

It's more like they are a sophisticated world modeling program that builds a world model (or approximate "bag of heuristics") modeling the state of the context provided and the kind of environment that produced it, and then synthesize that world model into extending the context one token at a time.

But the models have been found to be predicting further than one token at a time and have all sorts of wild internal mechanisms for how they are modeling text context, like building full board states for predicting board game moves in Othello-GPT or the number comparison helixes in Haiku 3.5.

The popular reductive "next token" rhetoric is pretty outdated at this point, and is kind of like saying that what a calculator is doing is just taking numbers correlating from button presses and displaying different numbers on a screen. While yes, technically correct, it's glossing over a lot of important complexity in between the two steps and that absence leads to an overall misleading explanation.

[–] kromem@lemmy.world 6 points 1 month ago

They don't have the same quirks in some cases, but do in others.

Part of the shared quirks are due to architecture similarities.

Like the "oh look they can't tell how many 'r's in strawberry" is due to how tokenizers work, and when when the tokenizer is slightly different, with one breaking it up into 'straw'+'berry' and another breaking it into 'str'+'aw'+'berry' it still leads to counting two tokens containing 'r's but inability to see the individual letters.

In other cases, it's because models that have been released influence other models through presence in updated training sets. Noticing how a lot of comments these days were written by ChatGPT ("it's not X — it's Y")? Well the volume of those comments have an impact on transformers being trained with data that includes them.

So the state of LLMs is this kind of flux between the idiosyncrasies that each model develops which in turn ends up in a training melting pot and sometimes passes on to new models and other times don't. Usually it's related to what's adaptive to the training filters, but it isn't always can often what gets picked up can be things piggybacking on what was adaptive (like if o3 was better at passing tests than 4o, maybe gpt-5 picks up other o3 tendencies unrelated to passing tests).

Though to me the differences are even more interesting than the similarities.

[–] kromem@lemmy.world 1 points 1 month ago

I'm a proponent and I definitely don't think it's impossible to make a probable case beyond a reasonable doubt.

And there are implications around it being the case which do change up how we might approach truth seeking.

Also, if you exist in a dream but don't exist outside of it, there's pretty significant philosophical stakes in the nature and scope of the dream. We've been too brainwashed by Plato's influence and the idea that "original = good" and "copy = bad."

There's a lot of things that can only exist by way of copies that can't exist for the original (i.e. closure recursion), so it's a weird remnant philosophical obsession.

All that said, I do get that it's a fairly uncomfortable notion for a lot of people.

[–] kromem@lemmy.world 3 points 1 month ago

They also identity the particular junction that seems the most likely to be an artifact of simulation if we're in one.

A game like No Man's Sky generates billions of planets using procedural generation with a continuous seed function that gets converted into discrete voxels for tracking stateful interactions.

The researchers are claiming that the complexity of where our universe's seemingly continuous gravitational behaviors meet up with the behaviors of continuous probabilities converting to discrete values when being interacted with in stateful ways is incompatible with being simulated.

But completely overlook that said complexity itself may be the byproduct of simulation, in line with independent emerging approaches in how we are simulating worlds.

[–] kromem@lemmy.world 2 points 1 month ago

Yes, just like Minecraft worlds are so antiquated given how they contain diamonds in deep layers that must have taken a billion years to form.

What a simulated world contains as its local timescale doesn't mean the actual non-local run time is the same.

It's quite possible to create a world that appears to be billions of years old but only booted up seconds ago.

[–] kromem@lemmy.world 1 points 1 month ago (1 children)

Have you bothered looking for evidence?

What makes you so sure that there's no evidence for it?

For example, a common trope we see in the simulated worlds we create are Easter eggs. Are you sure nothing like that exists in our own universe?

 

I often see a lot of people with outdated understanding of modern LLMs.

This is probably the best interpretability research to date, by the leading interpretability research team.

It's worth a read if you want a peek behind the curtain on modern models.

7
submitted 2 years ago* (last edited 2 years ago) by kromem@lemmy.world to c/technology@lemmy.world
 

I've been saying this for about a year since seeing the Othello GPT research, but it's nice to see more minds changing as the research builds up.

Edit: Because people aren't actually reading and just commenting based on the headline, a relevant part of the article:

New research may have intimations of an answer. A theory developed by Sanjeev Arora of Princeton University and Anirudh Goyal, a research scientist at Google DeepMind, suggests that the largest of today’s LLMs are not stochastic parrots. The authors argue that as these models get bigger and are trained on more data, they improve on individual language-related abilities and also develop new ones by combining skills in a manner that hints at understanding — combinations that were unlikely to exist in the training data.

This theoretical approach, which provides a mathematically provable argument for how and why an LLM can develop so many abilities, has convinced experts like Hinton, and others. And when Arora and his team tested some of its predictions, they found that these models behaved almost exactly as expected. From all accounts, they’ve made a strong case that the largest LLMs are not just parroting what they’ve seen before.

“[They] cannot be just mimicking what has been seen in the training data,” said Sébastien Bubeck, a mathematician and computer scientist at Microsoft Research who was not part of the work. “That’s the basic insight.”

view more: next ›