this post was submitted on 04 Dec 2023
1 points (100.0% liked)

LocalLLaMA

1 readers
1 users here now

Community to discuss about Llama, the family of large language models created by Meta AI.

founded 10 months ago
MODERATORS
 

What is everyone's experiences so far with DPO trained versions of their favorite models? Been messing around with different models and my two new favorite models are actually just the DPO versions of my previous favorite models (causalLM 14b and openhermes 2.5 7b). Links below for the models in question.

CausalLM 14B-DPO-alpha - GGUF: https://huggingface.co/tastypear/CausalLM-14B-DPO-alpha-GGUF

NeuralHermes 2.5 Mistral 7B - GGUF: https://huggingface.co/TheBloke/NeuralHermes-2.5-Mistral-7B-GGUF

The former runs at 30 t/s for me with koboldcpp-rocm on a 6900 XT, and the latter at 15 t/s, both at Q6K. I don't have a favorite between these two models, they seem to be better at different things and trade blows in all the logic + creative writing tasks I've tested them in, despite causalLM being a larger model. I'm looking forward to seeing what nousresearch/teknium and CausalLM are bringing next.

you are viewing a single comment's thread
view the rest of the comments
[–] 1monster90@alien.top 1 points 9 months ago (4 children)

Out of topic here but is it just me or open hermes somehow speaks in russian unprovoked frequently?

[–] VertexMachine@alien.top 1 points 9 months ago

Interesting. I'm using oobabooga and that never happened to me. I actually don't recall it ever outputting anything but English...

[–] ttkciar@alien.top 1 points 9 months ago

If you are using llama.cpp, you might want to give it a grammar which forces ASCII output.

[–] xadiant@alien.top 1 points 9 months ago

High repetition penalty? One model I merged suddenly started speaking Spanish in one summarisation task lol

[–] Lacono77@alien.top 1 points 9 months ago

I get polish or something