I use Runpod for all of my experiments. It’s less than half the price of AWS and is easy to work with.
LocalLLaMA
Community to discuss about Llama, the family of large language models created by Meta AI.
I find https://cloud.vast.ai to have best prices for personal stuff. Would not trust it with professional stuff though.
I use runpod for everything I can't do locally and I've been very happy with it. I initially chose it just because it was one of the cheapest, indeed way cheaper than the big 3, but I've had a good experience.
The main downside I know of runpod is that you can only run a container image, you can't have a full VM. but for most use cases I think this is really no big deal. if you want a generic sandbox for interactive experimentation, rather than to run an actual containerized app, you can just use the runpod pytorch image to get a starting point with cuda and pytorch and some other common stuff installed and then just ssh into it and do whatever. i.e. you don't necessarily have to bother with a more "normal" containerized deployment where you're writing something that runs unattended or exposes an API or whatever, writing a dockerfile etc
full disclosure my recent experiments are all testing different setups for inference with continuous batching, i'm personally not doing training or finetuning. but as far as I can tell runpod would be equally applicable for training and finetuning tasks
Together.AI has very fast inference speed and you pay by the token. 5M tokens of OpenHermes for a dollar.