this post was submitted on 09 Oct 2023
34 points (100.0% liked)

Futurology

2487 readers
7 users here now

founded 2 years ago
MODERATORS
top 11 comments
sorted by: hot top controversial new old
[–] Lugh@futurology.today 14 points 2 years ago (1 children)

One of the prevalent doomerist ideas about AI is that big tech will control it all in the future. Yet reality is behaving totally differently. Open-source AI seems only a few months behind big tech.

[–] conditional_soup@lemm.ee 5 points 2 years ago (1 children)

Which is why big tech is currently trying to convince world leadership that only they can be trusted to properly align powerful AI models. It's almost time to do the ol' ladder pull.

[–] SchizoDenji@lemm.ee 2 points 2 years ago

Corporations would love open source Ai models since it saves them the money that they'd have to pay to someone like openai.

[–] sudo22@lemmy.world 7 points 2 years ago (2 children)

Can this be easily self hosted?

[–] Speculater@lemmy.world 0 points 2 years ago (1 children)

The problem is most of these models need like a terabyte of VRAM... And consumers have about 8-24GB.

[–] sudo22@lemmy.world 2 points 2 years ago (1 children)
[–] Speculater@lemmy.world 4 points 2 years ago (1 children)

This specific one says it'll run on 24GB actually. But some are just crazy big.

[–] Lugh@futurology.today 2 points 2 years ago* (last edited 2 years ago) (1 children)

There are smaller models that can run on most laptops.

https://www.maginative.com/article/stability-ai-releases-stable-lm-3b-a-small-high-performance-language-model-for-smart-devices/

In benchmarks this looks like it is not far off Chat-GPT 3.5.

[–] BetaDoggo_@lemmy.world 1 points 2 years ago

It's not even close, less than half of 3.5's 85.5% in ARC. Some larger Open models are competitive in Hellaswag, TruthfulQA and MMLU but ARC is still a major struggle for small models.

3Bs are kind of pointless right now because the machines with processors capable of running them at a usable speed probably have enough memory to run a 7B anyway.

[–] BetaDoggo_@lemmy.world 0 points 2 years ago

It's a solid model but comparing to gpt4 immediately removes any credibility that they had.