Hoppss

joined 1 year ago
[–] Hoppss@alien.top 1 points 11 months ago

This is not true, I have split two separate LLM models partially across a 4090 and a 3080 and have had them both run inference at the same time.

This can be done in oobabooga's repo with just a little tinkering.