this post was submitted on 25 Nov 2023
1 points (100.0% liked)

LocalLLaMA

4 readers
4 users here now

Community to discuss about Llama, the family of large language models created by Meta AI.

founded 2 years ago
MODERATORS
 

So RWKV 7b v5 is 60% trained now, saw that multilingual parts are better than mistral now, and the english capabilities are close to mistral, except for hellaswag and arc, where its a little behind. all the benchmarks are on rwkv discor, and you can google the pro/cons of rwkv, though most of them are v4.

Thoughts?

you are viewing a single comment's thread
view the rest of the comments
[–] vasileer@alien.top 1 points 2 years ago (6 children)

I tested the 3B model and it looks good, especially the multilingual part (demo https://huggingface.co/spaces/BlinkDL/RWKV-Gradio-2)

[–] MoffKalast@alien.top 1 points 2 years ago (2 children)

Well it seems a lot better at Slovenian than LLamas or Mistral, especially for a 3B model, although it mostly just rambles about stuff that's vaguely related to the prompt and makes lots of grammatical mistakes. The 7B one ought to be interesting once it's done.

[–] vatsadev@alien.top 1 points 2 years ago (1 children)

Its trained on 100+ languages, the focus is multilingual

[–] alchemist1e9@alien.top 1 points 2 years ago

Will that make it a good translator? I remember seeing somewhere a 400+ language translation model but not an LLM somewhere. Wonder what the best many language open source fast high quality translation solutions might look like.

load more comments (3 replies)