because they are run by the borg (microsoft)
never think that ease is the only reason to do something privacy security
and overall control of your own domain are very good reasons.
another great reason local never says no.
because they are run by the borg (microsoft)
never think that ease is the only reason to do something privacy security
and overall control of your own domain are very good reasons.
another great reason local never says no.
ye im with ya on that multiple 3090's are the go unless your working massive models I think.
its really about the data curation and normalization.
think yi-34B they are getting results the same and better than 70B LLM's due
to the quality of there data.
work on the data and you will more than likely be happy with the results.
the training is really the fast part when you think of what is required to really nail down quality input.