this post was submitted on 08 Nov 2023
1 points (100.0% liked)

LocalLLaMA

1 readers
1 users here now

Community to discuss about Llama, the family of large language models created by Meta AI.

founded 10 months ago
MODERATORS
 

TL;DR:

Hey everyone, I am excited to share with you the first release of “DreamGen Opus”, an uncensored model that lets you write stories in collaborative fashion, but also works nicely for chat / (E)RP.

Specifically, it understands the following prompt syntax (yes, another one — please don’t hate :D):

(Description of the story, can also optionally include information about characters) 

...


(Instructions as you write the story, to guide the next few sentences / paragraphs)

You can find more details about prompting the model in the official prompting guide, including a few examples (like for chat / ERP).

The initial model is based on Mistral 7B, but Llama 2 70B version is in the works and if things go well, should be out within 2 weeks (training is quite slow :)).

The model is based on a custom dataset that has >1M tokens of instructed examples like the above, and order of magnitude more examples that are a bit less instructed.

How to try it out

The model should work great with any tool that supports the Mistral 7B base model. It will work well with oobabooga/text-generation-webui and many other tools. I like vLLM.

Using vLLM

  • Install vLLM following the instructions in the repo
  • Run python -u -m vllm.entrypoints.openai.api_server --host 0.0.0.0 --model dreamgen/opus-v0-7b

Using DreamGen.com website (free)

You can also try the model on dreamgen.com for free (but it requires a registration with email).

What’s next

I believe that for story telling & character creation it’s especially important to have access to the model weights, otherwise you run the risk of losing your plot or virtual companion (as already happened a few times before on various closed platforms that suddenly changed their rules or got shut down by their API provider). Hence DreamGen.

Here’s a high level overview of what I would like to do next under the DreamGen umbrella:

On the model side:

  • (Soon) Larger story models
  • Fine tune the model for even better character chat & roleplay
  • Longer context windows, at least for smaller models (8-16K depending on how experiments go)

On the application side, I am thinking about these features:

  • Character editor, chat & roleplay
  • Ability to share your stories privately & publicly (not sure about this one, to be honest :))
  • Image generation to go alongside with story generation & chat
  • API so that you can use the model more easily if you don’t have a GPU

For all of these, I would love your input! You can vote on the roadmap here.

For more updates, join the community server or follow updates on Twitter.

top 25 comments
sorted by: hot top controversial new old
[–] mcmoose1900@alien.top 1 points 10 months ago

but Llama 2 70B version is in the works

Might I suggest you use Yi-34B-200K instead? Or maybe later?

The problem I always have with storytelling models is slamming into the context limit, but Yi is already storytelling well out to 42K tokens for me, with just a basic Alpaca LoRA.

[–] trollsalot1234@alien.top 1 points 10 months ago

I have no idea how your model is but that prompting guide is probably the nicest one I've seen, so kudos on that.

[–] The_One_Who_Slays@alien.top 1 points 10 months ago

Sooo, how's the model?

[–] harrro@alien.top 1 points 10 months ago (2 children)

Any GGUF quantized download available?

[–] DreamGenX@alien.top 1 points 10 months ago

There was a bug on the website where the first time the "Continue" would not work if you did not refresh, should work now even though the editor is quite janky still, sorry for that :(

(can't wait for AI to take over React from me :P)

[–] mcmoose1900@alien.top 1 points 10 months ago

I was going to suggest you triain on Yi 34B 200K instead of Llama 70B, as my biggest issue with storytelling models is slamming into the context limit.

...But I just remembered that Yi has a stupid noncommercial license https://huggingface.co/01-ai/Yi-34B-200K/blob/main/LICENSE

Ugh. I hope a long context ~34B comes out that doesn't have such an ugly license.

[–] AbsorbingCrocodile@alien.top 1 points 10 months ago

Why only 7B?

[–] Revolutionalredstone@alien.top 1 points 10 months ago

Awesome l

Becoming your own AI company has never been easier 😊

[–] Dazzling_Ad1507@alien.top 1 points 10 months ago (1 children)

Very cool website and model!

[–] DreamGenX@alien.top 1 points 10 months ago
[–] deccan2008@alien.top 1 points 10 months ago (1 children)

Currently seems very expensive. Use of 7b models is effectively available for free in many places, Openrouter, Agnaistic, etc. Seems ridiculous that you don't get unlimited usage even with a subscriptions.

[–] DreamGenX@alien.top 1 points 10 months ago

I agree, I hope I can make things cheaper with better utilization. You have to consider that a single GPU is not used 100% the time, so there's a lot of waste. And due to lack of scale, I also do not get any special pricing on the GPUs. The more users, the closer the utilization will be to 100%, and the better GPU pricing. (For instance, I heard that on Google Cloud, enterprise customers can negotiate the on-demand GPU price down to the regular spot price for some of the GPUs)

[–] Proud-Point8137@alien.top 1 points 10 months ago

Is this the first fully uncensored mistral 7b?

[–] trollsalot1234@alien.top 1 points 10 months ago (2 children)

Alright, super technical review time: I got this running on the potato I connect to Reddit with, even though I usually only try gguf and only on days when the sun is shining and God seems happy. It made my Gtx 1070 ti cry (see I told you I would be technical!), but it worked. Then I altered a demo prompt, and it wrote me a story at about 1 token every 3 seconds where Little Red Riding Hood drank pee. So I'm giving this model a score of 8.6 dead babies, which is better than Tiefighter.

[–] DreamGenX@alien.top 1 points 10 months ago

Wow, amazing, thanks for giving it a try GGUF and other quants are coming, so your computer should have an easier time soon! :)

What's the maximum possible dead babies score? :D

[–] vitlaska@alien.top 1 points 10 months ago

Amazing. Reminds me of my favorite story testing prompt: [insert character] tricking Dr. Manhattan into drinking their piss at an Irish Pub. Can't wait to try it out with this one.

[–] Healthy_Cry_4861@alien.top 1 points 10 months ago

Looking forward to the release of 70b!

[–] Shaggy07tr@alien.top 1 points 10 months ago (1 children)

is the context length 2048?

[–] DreamGenX@alien.top 1 points 10 months ago

The training data had example of up to 4096 tokens. The model should also work beyond that, but I did not do a deep analysis of degradation.

[–] sharockys@alien.top 1 points 10 months ago (1 children)

Thank you for sharing! I am going to learn cooking with it.

[–] DreamGenX@alien.top 1 points 10 months ago

I hope it will be something tasty! :)

[–] erasels@alien.top 1 points 10 months ago (1 children)

This model is really good! By using the prompting template provided by the OP in the comments, I managed to get some really cool stories going. I haven't really tried stories with other models before, so maybe this is the norm for Mistral-based models, but I'm heavily impressed regardless.

Not getting any of the common blocker phrases once you ask for something even slightly bad to happen to a character is also pretty nice.

[–] DreamGenX@alien.top 1 points 10 months ago

Thank you so much for the kind feedback! If you have found some cool prompts, come share them with others on our discord.