this post was submitted on 13 Nov 2023
1 points (100.0% liked)

LocalLLaMA

1 readers
1 users here now

Community to discuss about Llama, the family of large language models created by Meta AI.

founded 10 months ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[โ€“] drplan@alien.top 1 points 10 months ago (1 children)

Perfect. Next please a chip that can do half the inference speed of an A100 with 15 Watts power.

[โ€“] MrTacobeans@alien.top 1 points 10 months ago

I don't think that will come from Nvidia. It's going to take in memory compute to get anywhere near that level of efficiency. First samples of these SOCs are no where near the memory requirements needed even for small models. These type of accelators will likely come from Intel/arm/risc/amd before Nvidia does it.