Wow, you guys are the best, could you also add estimated time for my run to start, thinking if i ll get something in meaningful time, but the mere fact things like this exist is great
Machine Learning
Community Rules:
- Be nice. No offensive behavior, insults or attacks: we encourage a diverse community in which members feel safe and have a voice.
- Make your post clear and comprehensive: posts that lack insight or effort will be removed. (ex: questions which are easily googled)
- Beginner or career related questions go elsewhere. This community is focused in discussion of research and new projects that advance the state-of-the-art.
- Limit self-promotion. Comments and posts should be first and foremost about topics of interest to ML observers and practitioners. Limited self-promotion is tolerated, but the sub is not here as merely a source for free advertisement. Such posts will be removed at the discretion of the mods.
Giving their gpu for free - this is some iq 200 stuff
Do you allow training of other sorts of models? I want to train a TTS model.
We support only large models (starting from 7B).
By 'training', I assume you mean fine-tuning or LoRA?
We only do full fine-tune.
Are you having good luck with adding knowledge to the model? I tried this with llama for a couple weeks when things were just getting going and I just could not find good hyperparameters for fine tuning. I was also doing Lora so...idk.
Same
From our experience, to get a very good results you need
-
High quality dataset. It's worth to spend more time on data cleaning. It's way better to have a smaller dataset with high quality points than a huge dataset with garbage.
-
You need to fully finetune it.
Don't leave us hanging, what does the cluster look like? (ignore if you're not allowed to share, but I'm a gigantic hardware nerd)
In terms of their capacity nothing crazy, Its probably a standard H100 or A100 cluster, 32 or 64 gpus
Why are you hiding who you are, and how many GPUs you have ... and if you have legal access to them?
What's with the tendency for software engineers to name their libraries after fundamental physics? As a physicist this always bothered me. I'll search for numerical algorithms for doing real physics.. and end up with some garbage blockchain app or a Rust crate that does nothing
I'm a bot, bleep, bloop. Someone has linked to this thread from another place on reddit:
- [/r/datascienceproject] Higgsfield.AI – Anyone can train Llama 70B or Mistral for free (r/MachineLearning)
^(If you follow any of the above links, please respect the rules of reddit and don't vote in the other threads.) ^(Info ^/ ^Contact)