Mission_Revolution94

joined 11 months ago
[–] Mission_Revolution94@alien.top 1 points 11 months ago

its really about the data curation and normalization.

think yi-34B they are getting results the same and better than 70B LLM's due

to the quality of there data.

work on the data and you will more than likely be happy with the results.

the training is really the fast part when you think of what is required to really nail down quality input.

[–] Mission_Revolution94@alien.top 1 points 11 months ago

because they are run by the borg (microsoft)

never think that ease is the only reason to do something privacy security

and overall control of your own domain are very good reasons.

another great reason local never says no.

[–] Mission_Revolution94@alien.top 1 points 11 months ago

ye im with ya on that multiple 3090's are the go unless your working massive models I think.