this post was submitted on 23 Nov 2023
1 points (100.0% liked)

LocalLLaMA

3 readers
1 users here now

Community to discuss about Llama, the family of large language models created by Meta AI.

founded 1 year ago
MODERATORS
 

Yi is a series of LLMs trained from scratch at 01.AI. The models have the same architecture of Llama, making them compatible with all the llama-based ecosystems. Just in November, they released

  • Base 6B and 34B models
  • Models with extended context of up to 200k tokens
  • Today, the Chat models

With the release, they are also releasing 4-bit quantized by AWQ and 8-bit quantized by GPTQ

Things to consider:

  • Llama compatible format, so you can use across a bunch of tools
  • License is not commercial unfortunately, but you can request commercial use and they are quite responsive
  • 34B is an amazing model size for consumer GPUs
  • Yi-34B is at the top of the OS Leaderboard, making it a very strong base model for a chat one
you are viewing a single comment's thread
view the rest of the comments
[–] Hatter_The_Mad@alien.top 1 points 11 months ago (2 children)

Can you give a example of such a model?

[–] a_beautiful_rhind@alien.top 1 points 11 months ago (2 children)
[–] azriel777@alien.top 1 points 11 months ago

For whatever reason, I keep getting memory errors with nous, but can run yi 34b fine. No idea what is wrong.

[–] reddithotel@alien.top 1 points 11 months ago (1 children)

I cannot load that one :(. Dolphin does work for me, but I cannot change the output writing style.

[–] a_beautiful_rhind@alien.top 1 points 11 months ago

Sucks, all the ones I d/l work so far but I'm using exl2.

Those are actually 2 different 34b chat models but there is a merge of them, nous-tess. They were the first that came to mind. If you search 34b there are others.

[–] SomeOddCodeGuy@alien.top 1 points 11 months ago

I started using the Tess Yi 34b 200k the other day and it's been fantastic. Of the 34bs I've tried so far, I'm really enjoying it the most for an AI assistant. I definitely recommend toying around with it.

If you do like me and use a Vicuna instruct, make sure to put a custom stop string of ""