This model is primarily recommended as a superior-to-Llama-2 baseline for additional finetuning,
According to the model, its not really supposed to compete with something like Vicuna. Sounds like they're trying to be an upgraded foundational model.
Community to discuss about Llama, the family of large language models created by Meta AI.
This model is primarily recommended as a superior-to-Llama-2 baseline for additional finetuning,
According to the model, its not really supposed to compete with something like Vicuna. Sounds like they're trying to be an upgraded foundational model.
What have you found it useful for? The model card is pretty vague.
its model average on the openllm leaderboard is 51.
Really nice, I had a dreamz we need to find a way to iterate over base models so every finetune is closer to sota :D
53 GB?
it's in FP32 rather than FP16
I really wonder who this TheBlok is. What a legend.
I can't speak to the quality of sequelbox/DaringFortitude but I can wholeheartedly recommend sequelbox/StellarBright. I have been using StellarBright in some experimental 70b model merges and it's phenomenal. I imagine 13b merges using DaringFortitude, or finetunes on top of it, would be quite good.
There is very little info.
It seems to be instruction finetuned, but what template? ChatML? There is no mention of anything. Posting it this way is pretty bad.