Herr_Drosselmeyer

joined 10 months ago
[–] Herr_Drosselmeyer@alien.top 1 points 9 months ago

Use 10 second clips of clean audio, no music, no background noise. I like to record samples from audiobooks. Free samples on Amazon recorded with audacity work well for me.

One thing to note, my install (an implementation for SillyTavern) somehow got corrupted, no idea how. It still worked but sounded way worse. Reinstall fixed that so maybe that's happening to you too.

[–] Herr_Drosselmeyer@alien.top 1 points 9 months ago

With a 3090 and sufficient system RAM, you can run 70b models but they'll be slow. About 1.5 tokens/second. Plus quite a bit of time for prompt ingestion. It's doable but not fun.

[–] Herr_Drosselmeyer@alien.top 1 points 9 months ago

Because a frightening amount of people still think Twitter matters.

[–] Herr_Drosselmeyer@alien.top 1 points 9 months ago

The models don't have memory per se, they just process the entirety of the context (i.e. the conversation) with each generation. As this becomes larger and more complex, models with less parameters struggle.

You can try to add certain instructions into the system prompt, such as "advance the story" but ultimately, more parameters means better grasp of the conversation. I haven't come across any model below an 8 bit 13b model that could keep a story together, so that's the minimum I go for when I want to RP.

As for the 70b's writing being less interesting, I'd say that's independent of the model capabilities and more down to style. Again, giving it instructions on how to write as well as example messages can help but it does somewhat come down to what it was trained on.

[–] Herr_Drosselmeyer@alien.top 1 points 9 months ago

It's a rule of thumb that yes, higher parameter at low quant beats lower parameter at high quant (or no quant) but take it with a grain of salt as you may still prefer a lower parameter model that's more tuned for the task you prefer.

[–] Herr_Drosselmeyer@alien.top 1 points 9 months ago (2 children)

The model, called Q* – and pronounced as “Q-Star” – was able to solve basic maths problems it had not seen before, according to the tech news site the Information, which added that the pace of development behind the system had alarmed some safety researchers.

Sound like a load of bullocks to me. How would anybody working in AI be "alarmed" by a model solving basic maths problems?

[–] Herr_Drosselmeyer@alien.top 1 points 9 months ago

Try just exllama2, no HF.

[–] Herr_Drosselmeyer@alien.top 1 points 9 months ago

I know but it's slowing down quite a bit at 32k already so I don't think it's worth pushing it further. But hey, even at just 16k it's four times what we usually get, so I'm not complaining.

[–] Herr_Drosselmeyer@alien.top 1 points 9 months ago (3 children)

With this particular model, I can crank it up to 32k if I enable " Use 8-bit cache to save VRAM" and that's as high as it can go in Oobabooga WebUI.

[–] Herr_Drosselmeyer@alien.top 1 points 9 months ago (11 children)

The base Yi can handle 200k. The version I used can do 48k (though I only tested 16k so far). Larger context size requires more VRAM.

The size that TheBloke like gives for GGUF is the minimum size at 0 context. As context increases, VRAM use increases.

 

I tried one yesterday for the first time and I came away very impressed. It's not great (yet) for RP but the basis seems very robust. For a laugh, I tried the usual reasoning tests and damn, it passed them all first try:

AI

How can I help you today?

You

There is a room with three murderers inside. Another person enters the room and shoots one of the three original occupants dead. Nobody else enters or leaves the room. How many murderers are there in the room?

AI

Four. Three original occupants who were already murderers and the fourth being the new person who entered the room and killed one of them.

You

I buy ten apples at the market. I give two of them to my friend Bob and two to my mother. Then, I go back to the market and buy three more apples. On my way home, I eat one apple. At home, I give my brother three bananas. How many apples do I have left. Explain your reasoning step by step.

AI

Let's break it down step-by-step:

You start with 10 apples.

You give away 2 apples to Bob (leaving you with 8 apples) and then give away another 2 apples to your mother (leaving you with 6 apples).

Next, you purchase 3 additional apples from the market, bringing your total up to 9 apples again.

While on your way home, you consume 1 apple, leaving you with 8 apples remaining.

Finally, when you reach home, you give your brother 3 bananas - but this doesn't affect the number of apples you have, since they are different fruits. So, you still have 8 apples left.

You

Sally has five brothers. Each of her brothers has one sister. How many sisters does Sally have. Explain youur reasoning step by step.

AI

Here's how we can reason through this problem:

Sally has five brothers.

It states that each of her brothers has one sister.

Since all these siblings belong to the same family, it means that Sally is the only sister mentioned for those five brothers.

Therefore, Sally has no other sisters besides herself.

So, Sally has 0 sisters

Like, seriously? Huge context, really fast (on my 3090 anyway) and this level of performance. I can't wait to see what people can do with this.

[–] Herr_Drosselmeyer@alien.top 1 points 10 months ago

I was hoping for a shakeup and all we got was an expensive game of musical chairs? Meh.

view more: next ›