So it was a perf test of a 1b token size model not the full 3.7T that get3 is trained with. I mean great. They are showing improvement but this is just a headline grabber they haven't done anything actually useful here.
this post was submitted on 09 Nov 2023
20 points (100.0% liked)
Futurology
0 readers
1 users here now
founded 1 year ago
MODERATORS
Just checking in to say they are still there - so many rascals showing off rigs these days