nsfw_throwitaway69

joined 1 year ago
[–] nsfw_throwitaway69@alien.top 1 points 11 months ago

Goliath wasn't fine-tuned at all, it's just a merge.

[–] nsfw_throwitaway69@alien.top 1 points 11 months ago (2 children)

None at all, it's just a merge. I'm not even really sure where to begin training it lol.

[–] nsfw_throwitaway69@alien.top 1 points 11 months ago

I don't think this is true. Goliath wasn't fine-tuned or trained at all and it outperforms every 70b I've ever used.

[–] nsfw_throwitaway69@alien.top 1 points 11 months ago (10 children)

Hi, I'm the creator of Venus-120b.

Venus has Synthia 1.5 mixed in with it, which as you noted performs pretty badly on RP. I'm currently working on a trimmed down version of Venus that has 100b parameters and I'm using SynthIA 1.2b for that, which I believe scored much better in oyur last RP tests. I'll probably also make a 1.1 version of Venus-120b that uses SynthIA 1.2b as well to see if that helps fix some of the issues with it.

[–] nsfw_throwitaway69@alien.top 1 points 11 months ago (1 children)

Venus-120b is actually a bit bigger than Goliath-120b. Venus has 140 layers and Goliath has 136 layers, so that would explain it.

[–] nsfw_throwitaway69@alien.top 1 points 11 months ago (3 children)

Crap, what's your setup? I tested it with a single 48GB card but if you're using 2x 24 then it might not work. I'll have to make a 2.8 bpw quant (or get someone else to do it) so that it'll work with card splitting.

[–] nsfw_throwitaway69@alien.top 1 points 11 months ago

Yeah I wanted a picture to go with the model and that's what stable diffusion spat out :D

And I haven't tried it for SFW stuff but my guess is that it would work fine.

[–] nsfw_throwitaway69@alien.top 1 points 11 months ago (2 children)

πŸ€” How are you trying to load it? I tested both quants in text-generation-webui and they worked fine for me. I used exllama2_hf to load it

[–] nsfw_throwitaway69@alien.top 1 points 11 months ago (2 children)

Hard to say. Try it out and let me know!

[–] nsfw_throwitaway69@alien.top 1 points 11 months ago

Try it out and let me know! I included Nous-Hermes in the merge because I've found it to be one of the best roleplaying models that doesn't hallucinate too much. However, Nous-Hermes also tends to lack a bit in terms of the prose it writes, from my experience. I was hoping to get something that's coherent most of the time and creative.

[–] nsfw_throwitaway69@alien.top 1 points 11 months ago

Thanks! I'm eager to see the results :)

 

Hi everyone, I'd like to share something that I've been working on for the past few days: https://huggingface.co/nsfwthrowitaway69/Venus-120b-v1.0

This model is the result of interleaving layers from three different models: Euryale-1.3-L2-70B, Nous-Hermes-Llama2-70b, and SynthIA-70B-v1.5, resulting in a model that it larger than any of the three used for the merge. I have branches on the repo for exl2 quants at 3.0 and 4.85 bpw, which will allow the model to run in 48GB or 80GB of vram, respectively.

I love using LLMs for RPs and ERPs and so my goal was to create something similar to Goliath, which is honestly the best roleplay model I've ever used. I've done some initial testing with it and so far the results seem encouraging. I'd love to get some feedback on this from the community! Going forward, my plan is to do more experiments with merging models together, possibly even going even larger than 120b parameters to see where the gains stop.

view more: next β€Ί