silenceimpaired

joined 1 year ago
[–] silenceimpaired@alien.top 1 points 11 months ago

Exciting and worrying… I have gone to great efforts to use safetensors… I would have to see every model packaged in executable format… but then again I have seen comments about llama.cpp behavior changing for the same model and settings (not sure if it is true but that could be bad)

[–] silenceimpaired@alien.top 1 points 11 months ago (1 children)

Well then… Thanks! I’ll use llama.cpp and be happy. Glad to hear llamacpp_hf is crazy and not me. Which tool do you prefer outside of Oobabooga?

[–] silenceimpaired@alien.top 1 points 11 months ago (1 children)

It’s only been a day but have you changed? I find this model misspells a lot with the gguf i downloaded.

[–] silenceimpaired@alien.top 1 points 11 months ago (3 children)

So helpful… but Yi and llamacpp_hf just falls apart for me… complete gibberish on Oobabooga. Exl hf … fine. Llama.cpp fine… Min-P is there and I can apparently use it but temperature last is missing :/

[–] silenceimpaired@alien.top 1 points 11 months ago

I’m on Pop, lol. I could get it to compile, but I must have missed a step for nvidia acceleration

[–] silenceimpaired@alien.top 1 points 11 months ago

It is… but koboldcpp doesn’t have a executable for me to run :/

[–] silenceimpaired@alien.top 1 points 11 months ago (4 children)

I could never get up and running on Linux with Nvidia. I used Kobold on Windows, but boy is it painful on Linux.

[–] silenceimpaired@alien.top 1 points 1 year ago

Please try 70b down to ~30b with a llama 2 model. Thanks!