this post was submitted on 26 Nov 2023
1 points (100.0% liked)

LocalLLaMA

1 readers
1 users here now

Community to discuss about Llama, the family of large language models created by Meta AI.

founded 10 months ago
MODERATORS
 

I get 20 t/s with a 70B 2.5bpw model, but this is only 47% of the theoretical maximum of 3090.

In comparison, the benchmarks on the exl2 github homepage show 35 t/s, which is 76% the theoretical maximum of 4090.

The bandwidth differences between the two GPUs aren't huge, 4090 is only 7-8% higher.

Why? Does anyone else have a similar 20 t/s ? I don't think my cpu performance is the issue.

The benchmarks also show ~85% utilization on 34B on 4bpw (normal models)

you are viewing a single comment's thread
view the rest of the comments
[–] brobruh211@alien.top 1 points 9 months ago

Thanks for the detailed answer! Ubuntu does seem to be much more memory-efficient compared to Windows. However, the problem just fixed itself seemingly overnight. Now I'm not running into out of memory errors. 8-bit cache is a godsend for vram efficiency.