this post was submitted on 27 Oct 2023
1 points (100.0% liked)

Machine Learning

1 readers
1 users here now

Community Rules:

founded 10 months ago
MODERATORS
 

What are the benefits of using an H100 over an A100 (both at 80 GB and both using FP16) for LLM inference?

โ€‹

Seeing the datasheet for both GPUS, the H100 has twice the max flops, but they have almost the same memory bandwidth (2000 GB/sec). As memory latency dominates inference, I wonder what benefits the H100 has. One benefit could, of course, be the ability to use FP8 (which is extremely useful), but I'm interested in the difference in the hardware specs in this question.

you are viewing a single comment's thread
view the rest of the comments
[โ€“] SnooHesitations8849@alien.top 1 points 10 months ago

H100 and A100 are best for training. H100 is optimized for lower precision (8/16 bits) and optimized for transformer. A100 is still very good but not that much. A100 is still very GPU-like. Wwhile H100 is a transformer-accelerator.

Using them for inference is not the best econ-friendly though.