this post was submitted on 15 Feb 2024
151 points (93.6% liked)

Technology

59377 readers
5196 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
 

“In 10 years, computers will be doing this a million times faster.” The head of Nvidia does not believe that there is a need to invest trillions of dollars in the production of chips for AI::Despite the fact that Nvidia is now almost the main beneficiary of the growing interest in AI, the head of the company, Jensen Huang, does not believe that

you are viewing a single comment's thread
view the rest of the comments
[–] Buffalox@lemmy.world 8 points 9 months ago* (last edited 9 months ago) (1 children)

It requires 4X speed increase every year, production quality scale can't provide even close to half of that, maybe 25%, then another 25% from design, and regarding increasing die sizes they are already close to the end. So the only way to get from 150% to 400% per year is by using multi chip designs, meaning they will have to use 2.5x more chips per year. so the multi chip package in 10 years will probably have to have almost 10,000 chips! All of them bleeding edge!!!

The H200 is estimated to cost $40K, the future 10 year chip will be more like $40 million. Or maybe more like impossible to achieve.

[–] agent_flounder@lemmy.world 2 points 9 months ago (1 children)

If chips = cpus, here, then I imagine that will hit a limit also (Amdahl's law).

[–] Buffalox@lemmy.world 4 points 9 months ago (1 children)

A chip is also called a die, it's the piece cut out from the wafer, which is then packaged onto a chip package.
Since traditionally there were always 1 chip per chip package, the 2 words were used almost synonymously.
I this case it's basically GPU chips, which AFAIK AMD has already figured out how to use in multi chip packages. Meaning one package contains multiple chips that work "almost" as well as a single chip of similar size.

The advantage of multichip packages are obvious, production costs are way lower because smaller dies causes lower percentage of flawed dies, and allows for better binning of higher end parts.
Additionally it allows designs of way more complex packages, than would be possible with monolithic chips. This is the reason AMD has been taking marketshare in server markets from Intel. Because Intel has not been able to match the multichip design AMD introduced with Epyc in 2016/17, which originally was 4 Ryzen chiplets/chips/dies packaged together as one big 32 core server chip. Where the biggest Intel could make was 28 cores.

But packaging almost 10000 GPU chips together is completely different, and I don't think that will be relevant within 10 years.

Amdahls law however is part obvious and part bullshit. Everything your mind is able to do semi efficiently, can be multithreaded, it is very few things that can't.
Amdahls law is basically irrelevant with regard to AI, as AI has a lot of patten recognition, and pattern recognition is perfect for multi threading.

[–] TheGrandNagus@lemmy.world 3 points 9 months ago

And to add: currently TSMC nodes have a reticle limit of 858mm². I.e. that's the largest chips you can make on their wafers. Then in the real world you do it slightly below that.

Future nodes are reducing this to the 350-450mm² range.

High end GPUs/HPC cards basically have to go to multi-die, even in the fantasy world of 100% perfect yields.