Mission_Revolution94

joined 2 years ago

its really about the data curation and normalization.

think yi-34B they are getting results the same and better than 70B LLM's due

to the quality of there data.

work on the data and you will more than likely be happy with the results.

the training is really the fast part when you think of what is required to really nail down quality input.

because they are run by the borg (microsoft)

never think that ease is the only reason to do something privacy security

and overall control of your own domain are very good reasons.

another great reason local never says no.

ye im with ya on that multiple 3090's are the go unless your working massive models I think.