OC2608

joined 10 months ago
[โ€“] OC2608@alien.top 1 points 9 months ago (1 children)

How to earn VC money 101: "Beats GPT-4!"

And voila! you're rich now.

 

It's no secret that many language models and fine-tunes are trained using datasets, many of them are made using GPT models. The problem arises when many "GPT-isms" end up in the dataset. And I am not only referring to the typical expressions like "however, it's important to...", "I understand your desire to...", but I am also referring to the structure of the outputs in the model's responses. ChatGPT (GPT models in general) tend to have a very predictable structure when in its "soulless assistant" mode, which makes it very easy to say "this is very GPT-like".

What do you think about this? Oh, and by the way, forgive my English.

[โ€“] OC2608@alien.top 1 points 10 months ago

I used to use Text Generation Web UI, but I changed to KoboldCpp because it's more lightweight. Besides, I realized I didn't use all the features of the textgen UI. KoboldCpp as the backend and SillyTavern as the frontend when I want to chat. KoboldCpp alone when I want to play with models by creating stories or something.