this post was submitted on 24 Apr 2024
354 points (96.1% liked)

Technology

59596 readers
3946 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
 

It started with notebooks, but that wasn’t the master plan.

you are viewing a single comment's thread
view the rest of the comments
[–] bitfucker@programming.dev 4 points 7 months ago (1 children)

I mean, what is the difference between the current SoC and the soldered CPU? Sure you can save on upgrading RAM, but then what else? Especially if the SoC has PCIe. They can make a daughter board for the SoC to make it simpler to upgrade if they want, alà pi compute module.

[–] Dudewitbow@lemmy.zip 3 points 7 months ago* (last edited 7 months ago) (1 children)

its not that simple. high performamce parts are high performamce because the devices that need the fastest speeds have the shortest traces from CPU to said device. its for instance, why the ram slots, and the fastest m.2/slot as well as pci-e lanes are nearest to the cpu, else youd have to resort to adding a south bridge.

the pi compute module works that way because the ram is already on board making it not a problem, and latency to whatever it gets mounted on isnt of highest priority for performance.

its why sodimm for instance has hit a peak speed limit, while lpdde hasnt, and why dell pitched the camm form factor for ram. distance of components to the cpu and its stability is cruicial for performance.

[–] bitfucker@programming.dev 1 points 7 months ago (1 children)

What I am saying is that the current soldered CPU approach on laptop space is not that different from switching an ARM SoC on a daughter board. The only difference is that you cannot change RAM. Maybe that too will change as you said with CAMM standard. Next is that some SBC already supported PCIe for external M.2 storage, so you can theoretically hook up a removable GPU there.

Now, what to do with the old SoC daughter board? The same as with the old framework motherboard. You can repurpose it as another computer.

The point is, framework repairability comes not only from part swapping, but also the promise of providing schematic for board level repair. They can totally make ARM based laptops with SoC repairable if they wanted to. But I suspect they will not (at least in the near future) since there is a lot to do for them.

[–] Dudewitbow@lemmy.zip 1 points 7 months ago

m.2 to gpu isn't completely foreign nor new, but less practical than more recent standards like Occulink. the problem, specifically with the lower end model in particular, is using 4/8 pci-e lanes for a port that not everyone is going to use is a waste of the already limited amount of pci-e lanes available to the user because of the CPU choice. hence, it makes sense to keep 1 users with the side option to using usb4/thunderbolt gpu docks