this post was submitted on 13 Nov 2023
1 points (100.0% liked)
LocalLLaMA
3 readers
1 users here now
Community to discuss about Llama, the family of large language models created by Meta AI.
founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
I was thinking of distributed MoEs as well.
Question I have is how do you route queries? I don't know how to do that if all the Es are in the same cluster let alone distrivuted.
I don’t know if there’s much value there when LORA’s are easily portable — you can just select the right lora as needed. One base model instance on one machine, many potential experts. This has been demonstrated.
yeah its a work in progress. Its not trivial to setup . it's easy to imagine a way it could be done , but it all has to be built, tested, refined.
llama cpp is out there, I am a c++ person but I dont have deep experience with LLMs (how to fine tune etc) generally and have other projects in progress. but if you look around in the usual places with some search terms you'll find the attempts in progress, and they probably could use volunteers.
my aspirations are more toward the vision side, I'm a graphics person and need to get on with producing synthetic data or something