this post was submitted on 22 Nov 2023
1 points (100.0% liked)

LocalLLaMA

3 readers
1 users here now

Community to discuss about Llama, the family of large language models created by Meta AI.

founded 1 year ago
MODERATORS
 

I've been in need of a dedicated training rig lately and so I've been looking at gpus. For context I'm already running training on a 16gb 3080ti laptop and inference on a 16gb 4060ti. Both are really just fine for 13B models.

When I'm looking at cards though it appears I could buy nearly 4 more 16gb 4060ti cards for the price of a 24gb 4090.

I understand that the 4090 is potentially 2-3 times faster based on benchmarks, but does this actually translate to improved Llama speeds? Would it even be viable to go for double 4060ti's instead?

Currently I'm standardized on 16gb/13B/4bit but I'd love to push beyond that, have more vram for training, etc. What are my options?

no comments (yet)
sorted by: hot top controversial new old
there doesn't seem to be anything here