this post was submitted on 28 Oct 2023
1 points (100.0% liked)

Machine Learning

1 readers
1 users here now

Community Rules:

founded 11 months ago
MODERATORS
 

I wanted to share some exciting news from the GPU world that could potentially change the game for LLM inference. AMD has been making significant strides in LLM inference, thanks to the porting of vLLM to ROCm 5.6. You can find the code implementation on GitHub.

The result? AMD's MI210 now almost matches Nvidia's A100 in LLM inference performance. This is a significant development, as it could make AMD a more viable option for LLM inference tasks, which traditionally have been dominated by Nvidia.

For those interested in the technical details, I recommend checking out this EmbeddedLLM Blog Post.

I'm curious to hear your thoughts on this. Anyone manage to run it on RX 7900 XTX?

https://preview.redd.it/rn7n29yxpuwb1.png?width=600&format=png&auto=webp&s=bdbac0d2b34d6f43a03503bbf72b446190248789

top 7 comments
sorted by: hot top controversial new old
[–] killver@alien.top 1 points 10 months ago

Is there some good cloud host for getting AMD GPUs?

[–] Booonishment@alien.top 1 points 10 months ago

Well now I know what I’m doing with my weekend. Thanks for sharing! Hopefully I can report back some xtx performance numbers.

[–] diamond_jackie07@alien.top 1 points 10 months ago (1 children)

I tried on this config - Ryzen 9 7950x MI210. I got this result Throughput: 129 requests/min, 1028.89 tokens/s on llama2-7b. Which is even better than the performance they cite on the post

[–] diamond_jackie07@alien.top 1 points 10 months ago

Will report back on 13b performance ASAP

[–] mrpoops@alien.top 1 points 10 months ago (1 children)

Will this run on a ryzen with Radeon built in?

If so, can’t you build a ryzen machine with like 128gb of ram and dedicate nearly all of it to video?

[–] TotesMessenger@alien.top 1 points 10 months ago

I'm a bot, bleep, bloop. Someone has linked to this thread from another place on reddit:

 ^(If you follow any of the above links, please respect the rules of reddit and don't vote in the other threads.) ^(Info ^/ ^Contact)