bot-333

joined 1 year ago
[–] bot-333@alien.top 1 points 11 months ago

I guess they might open source the 600B one? They have different names, so maybe different training approaches.

[–] bot-333@alien.top 1 points 11 months ago (2 children)

I think they changed it to it’s still an experiment and they are finishing evaluations to better understand the model.

[–] bot-333@alien.top 1 points 11 months ago

There are SO many models "bullshitting through some benchmarks or some other shenanigans" that I'm cooking my own benchmark system LOL.

[–] bot-333@alien.top 1 points 11 months ago (6 children)

"New RLAIF Finetuned 7b Model" Interesting. "beats Openchat 3.5" Nice! "and comes close to GPT-4" Bruh.

[–] bot-333@alien.top 1 points 11 months ago

Another day, another people trying to make a whole story based on two tokens.

[–] bot-333@alien.top 1 points 11 months ago (2 children)

Not sure if self promotion here is allowed. I found my own IS-LM 3B to be the most coherent, verbose, and factual/correct 3B I've tried. IMO it's better than Rocket 3B, but it scores worse in benchmarks. I suspect a contamination in Rocket 3B.

[–] bot-333@alien.top 1 points 11 months ago

Can you try my new IS-LM? GGUF: https://huggingface.co/UmbrellaCorp/IS-LM-3B_GGUF. I found it really good. Thanks.

[–] bot-333@alien.top 1 points 11 months ago

I suggest you to try IS-LM 3B.

[–] bot-333@alien.top 1 points 11 months ago (1 children)

Are you using the correct prompt template?

[–] bot-333@alien.top 1 points 11 months ago

I think I need to remind people of the benchmarks used, MT-Bench and AlpacaEval are terrible benchmarks.

 

Previously I made a post about my new model, IS-LM 3B. If you want more information of the model, read the post here.

As the creator of this model, I noticed that this model is extremely good for a 3B in economic tasks(As it is trained on it.), and surprising other tasks too. It follows instructions well, is VERY verbose, and is very accurate compared to other 3B models and even LLaMA 1 7B. Though it is bad at casual chats/roleplay, but that was expected because it was not trained on those. I am very surprised and impressed, and here's some example conversations(Warning: a LOT of text.): https://pastebin.com/2BZC7kZg

I did NOT cherrypick those, these are almost all of the proper tests I've done to the model. It is very consistent at generating those type of responses, if you don't believe me, try it out yourself. Also note that all these are seperate conversations, as I found that this model isn't the best at multi-turn conversations.

Obviously not all of them are correct or the best, but it does show very good capabilities and verbosity. This model is again, extremely good for a 3B, even outperforming a lot of the SOTA LLaMA 1 7B models IMO. If you are interested, I HIGHLY recommend you trying it out as it can be very easily ran.

 

Just if you are curious/confused of the name, the IS-LM model is basically "a two-dimensional macroeconomic tool that shows the relationship between interest rates and assets market"(Source.). I thought it's a creative name because IS-LM is both about economics and have "LM" in its name.

Anyways, I released IS-LM 3B. This model is fine-tuned on economics.

Details:

IS-LM 3B is StableLM 3B 4E1T(Licensed under CC BY-SA 4.0.) instruction tuned on DataForge Economics for 3 epochs with QLoRA(2305.14314).

Prompt template:

USER: {prompt}
ASSISTANT:
[–] bot-333@alien.top 1 points 11 months ago (1 children)

I see that their distilled model is much worse than StableLM 3E1T, so the finetuning improved a lot. Unfortunately they didn’t release the datasets(Would that still be considered Open Source?). Also I’m pretty sure my StableLM finetunes are better in the Open LLM Benchmarks, they just don’t allow StableLM models to be submitted.

 

If you don't know what those are, refer to these two reddit posts about Marx 3B V3 and Akins 3B, the unquantized model weights are available at HuggingFace. Link to Marx 3B V3 and Akins 3B.

As the StableLM support for llama.cpp has just been recently, u/The-Bloke(Thank you so much!) quantized my StableLM models to GGUF as a lot of people are wanting to try the model in GGUF, you can find the GGUF conversion for Marx 3B V3 and Akins 3B. Again, credit to u/The-Bloke for quantizing the model, thank you!

By the way, I don't know what dataset to finetune on right now. If you know a good dataset, let me know and I will look into it. Though I could probably only finetune on datasets below 5k conversations, maybe 10k.

 

Today I released Akins 3B.

Akins 3B is StableLM 3B 4E1T(Licensed under CC BY-SA 4.0.) instruction tuned on Claude Multiround Chat 1K for 2 epochs with QLoRA(2305.14314).

Prompt template:

USER: {prompt}
ASSISTANT:

From my current limited testing, this model is really impressive for a 3B. Example output:

Turbodurp, the wind-blown,
With hair as wild as the sea,
He walks the land with a swagger,
And a smile that could cheer you.

Turbodurp, the wise one,
Who sees the world with wonder,
He sees beauty in everything,
And joy in a simple sparrow.

Turbodurp, the joker,
Who laughs at the world's foibles,
He sees the humor in life's journey,
And finds joy in a simple toad.

Turbodurp, the dreamer,
Who sees the world in a dream,
He sees possibilities everywhere,
And hopes for a better scheme.

Turbodurp, the lover of life,
Who lives each day to the fullest,
He laughs at the world's troubles,
And finds joy in a simple cup of tea.

Turbodurp, the friend of all,
Who sees the good in everyone,
He lifts your spirits with his laughter,
And shares his dreams with you.

Turbodurp, the joyful one,
Who finds joy in simple things,
He sees the beauty in nature,
And finds joy in a simple ring.

Turbodurp, the one who loves,
Who loves with a heart of gold,
He loves with a love that endures,
And finds joy in a simple fold.
view more: next ›