Any details on what max context sizes are usable?
LocalLLaMA
Community to discuss about Llama, the family of large language models created by Meta AI.
I think I need to remind people of the benchmarks used, MT-Bench and AlpacaEval are terrible benchmarks.
As fan of the character, I approve ๐
Oh wow, this seems almost too good to be true
Woooooooow!
This smells like leftovers...
We've been having "pretraining on the test set" for weeks and I'm craving something else.
I think "The Bloke" takes requests for GUFF conversions. Might want to check hugging face.
๐ Training Data: We've amalgamated multiple public datasets to ensure a comprehensive and diverse training base. This approach equips Rocket-3B with a wide-ranging understanding and response capability.
We've amalgamated multiple public benchmark answers to ensure a contaminated and diverse training base.
Looking forward to trying this when some GGUF's are available.
Seems this model has a problem and not loading.
It was recently fixed then.
Finally, I can integrate AI to my arduino project and build my own version of BB-8
Tried gguf format of this model from huggingface and they just wont load.
I tried both GGUF models currently on HF. Same result.
Curious to try this out when it's working!
Same, even the model from the bloke that was released hours ago wouldn't work :-(
The latest version of KoboldCpp v1.50.1 now loads this model properly.
๐ฉโ๐ป Chat format: Rocket-3B follows the ChatML format.
From the README and the tokenizer.json it looks like it's using a textual representation of ChatML on top of StableLM's format. Just in case this trips anyone up.