this post was submitted on 28 Nov 2023
1 points (100.0% liked)

LocalLLaMA

3 readers
1 users here now

Community to discuss about Llama, the family of large language models created by Meta AI.

founded 1 year ago
MODERATORS
 

If cpu processing is slow af, and gpu takes $$$$ to get enough memory for larger models; I am wondering if an APU could deliver some of that gpu speed, but using cheaper ram to get the larger models in memory; with 128gb of ram, that’s the equivalent of 6x 30/4090s, without allowing for overhead at least!

Wondering if anyone has got any current apu benchmarks vs cpu/gpu? Do you know if the GPU side of APU architecture can be used to get an increase over traditional CPU results?

I’ve been seeing a lot of claims that the ryzen 8000 series is going to be competing with low end Gpus, some people think all the way up to 3060.

If it’s possible to do, it might be the new best way to get large models working for cheap?

you are viewing a single comment's thread
view the rest of the comments
[–] MINIMAN10001@alien.top 1 points 11 months ago (1 children)

It's not that CPUs are slow it's that typically RAM that the CPU is connected to is slow.

That's why unified memory is fast it's just faster and connected to the CPU.

[–] rarted_tarp@alien.top 1 points 11 months ago

The UMA has a lot more to do with the speed than distance, and GPU has a much different architecture and memory access patterns than a CPU.