Featureless_Bug

joined 1 year ago
[โ€“] Featureless_Bug@alien.top 1 points 11 months ago

I mean, Falcon 40 B with lora can easily be trained on 2x A100 with lora (even llama 70b can be trained on just 2x A100). But maybe accelerate is doing something stupid - in my experience, both deepspeed and accelerate are very slow and require way too much memory compared to manual gpu distribution strategy.

[โ€“] Featureless_Bug@alien.top 1 points 1 year ago (1 children)

Macbook Air is not money well spent. Workstation with 4090, and A100 in the cloud on the money you saved