this post was submitted on 17 Nov 2023
1 points (100.0% liked)

LocalLLaMA

1 readers
1 users here now

Community to discuss about Llama, the family of large language models created by Meta AI.

founded 10 months ago
MODERATORS
 

I've been encountering a repetition issue with models like Goliath 120b and Xwin 70b on Sillytavern + OpenRouter.While I understand that changing models can have a significant impact, I'm puzzled by the repetition problem.Despite my efforts to find online resources for correct settings, my searches for Auroboros 70b, Xwin 70b, Lzlb 70b, and others have been in vain.

I came across posts on this subreddit addressing similar concerns, but unfortunately, they lacked solutions.One suggestion was to "use the shortwave preset," but it seems to be nonexistent.Unsure of what I might be overlooking, I'm reaching out here for help.The 120b model should theoretically outperform the 7b/13b models, but I suspect there's a configuration issue.

If anyone could provide insights or share the correct settings for these models, it would greatly help not only me but also future users facing the same issue.Let's compile a comprehensive guide here so that anyone searching the internet for a solution can find this post and get the answers they need.Thank you in advance for your assistance!

PS: mythomax 13B seems to be the best model because it's the only one that actually works...

you are viewing a single comment's thread
view the rest of the comments
[–] a_beautiful_rhind@alien.top 1 points 10 months ago

I'm loving min_P + dynamic temperature. Feels like the only sampler I need.

These are better than shortwave, which was good in itself.

Another one to try is mirostat.

See what openrouter lets you use out of that.