Significant_Focus134

joined 1 year ago
[–] Significant_Focus134@alien.top 1 points 11 months ago

Thanks! For the record, that version is very under-trained. Today I started to train on much bigger dataset (50k entries) that is mostly built from the wikipedia.

 

Warning, this is still work in progress.

https://huggingface.co/piotr-ai/polanka-7b-v0.1

First version of 7b Polish LLM finetuned using custom data in Polish language.

As a base model I used uncensored https://huggingface.co/ehartford/dolphin-2.1-mistral-7b so Dolphin "personality" should also be there.

It was trained using 4K context in ChatML format. All done on a single 4090 for multiple days.