OldPin8654

joined 10 months ago
[–] OldPin8654@alien.top 1 points 10 months ago

Winter has enveloped us in its chilly embrace. In my quest for warmth, I realized I needed a heater. But then, a memory dawned on me – I had bought one before! It was during those long nights of training a model with a 100k dataset, which made the room toastier. Now, thanks to that, everyone in this house can enjoy a peaceful and warm winter.

[–] OldPin8654@alien.top 1 points 10 months ago (3 children)

Yeah, it is not my money but still stressful

[–] OldPin8654@alien.top 1 points 10 months ago

Not living in the US atm but no reputable sellers neither here 😂

[–] OldPin8654@alien.top 1 points 10 months ago

Yes! Put more money in it, the company!!!

[–] OldPin8654@alien.top 1 points 10 months ago (1 children)

I will talk to my boss for more money 😆

 

Hi all, I need a help from all of you. I am going to buy H100s for training LLMs. Currently for fine-tuning 70b models but later we may consider pre-training larger models too. H100s looks more promising than A100s considering its power of FP8 support, so I asked quotes from multiple vendors. And then, realized there are too many options!

  1. DGX - 8x H100, much more expensive than other options but they say its performance is worth it.

  2. Buy PCI-E H100 cards and a Supermicro machine - from 2x upto 8x, looks cost effective.

2.a. some vendors offered a combination with NVLinks. Some says 1 link is needed for 2 and some says 3 links are needed for 2.

  1. H100 NVL - no idea what the difference is compared to the PCI-E with NVLinks but looks like they are newly introduced ones.

  2. Some other options, like a custom build made by the vendors.

Any BEST PRACTICE I can take a look to make a decision? Any advice from experts here who suffered a similar situation already? Thanks in advance 🙏