this post was submitted on 28 Nov 2023
1 points (100.0% liked)
LocalLLaMA
3 readers
1 users here now
Community to discuss about Llama, the family of large language models created by Meta AI.
founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
We need some 4090s with 500gb VRAM modified in China if possible.
the devs mentioned that the 600B model takes about 1,3TB space alone..
Make it 0.01bpm quantized and you will fit in good ol' 3090.