GoofAckYoorsElf

joined 2 years ago
[–] GoofAckYoorsElf@alien.top 1 points 2 years ago (1 children)

that 64gb of RAM is cutting it pretty close

Holy crap...

[–] GoofAckYoorsElf@alien.top 1 points 2 years ago

Mistral-Hermes-2.5-7b-8bit

I've tried that one. It is... strange.

[–] GoofAckYoorsElf@alien.top 1 points 2 years ago (1 children)

nous-capybara-34b

I haven't been able to use that with my 3090Ti yet. I tried TheBloke's GPTQ and GGUF (4bit) versions. The first runs into memory issues, the second, loaded with llama.cpp (which it seems to be configured on) loads, but is excruciatingly slow (like 0.07t/sec).

I must admit that I am a complete noob regarding all the different variants and model loaders.

[–] GoofAckYoorsElf@alien.top 1 points 2 years ago (3 children)

I've been going with WizardLM-33B-V1.0-Uncensored-GPTQ for a while and it's okay. Is Nous-Capybara-34b better?

[–] GoofAckYoorsElf@alien.top 1 points 2 years ago

Quite some the stuff that commercial/corporate models won't let me do and which I wouldn't do even if they let me. Private stuff. Yes, NSFW can of course be a part of it.

Furthermore, things where I think the commercial/corporate models are too expensive (no, I have not checked my power bill yet...).