upvote
> This is a wild and very wrong take.

That's what happens when you quote only part of a statement. Taken in context, it was referring to a very real decline in expansion cards. Now that NICs (for WiFi) and SSDs have been moved into their own compact specialized slots, and Ethernet and audio have been standard integrated onto the motherboard itself for decades, the regular PCIe slots are vestigial. They simply are not widely used anymore for expanding a PC with a variety of peripherals (that era was already mostly over by the transition from 32-bit PCIe to PCIe).

Across all desktop PCs, the most common number of slots filled is one (a single GPU), and the average is surely less than one (systems using zero slots and relying on integrated graphics must greatly outnumber systems using more than one slot).

Even GPUs themselves are a horrible argument in favor of PCIe slots. The form factor is wildly unsuitable for a high-power compute accelerator, because it's ultimately derived from a 1980s form factor that prioritized total PCB area above all else, and made zero provisions for cards needing a heatsink and fan(s).

reply
> Ethernet and audio have been standard integrated onto the motherboard itself for decades

Unless the one it comes with isn't as fast as the one you want, or they didn't integrate one at all, or you need more than one.

> Across all desktop PCs, the most common number of slots filled is one (a single GPU), and the average is surely less than one (systems using zero slots and relying on integrated graphics must greatly outnumber systems using more than one slot).

There is an advantage in having an empty slot because then you can put something in it.

Your SSD gets full, do you want to buy one which is twice as big and then pay twice as much and screw around transferring everything, or do you want to just add a second one? But then you need an empty slot.

You bought a machine with an iGPU and the CPU is fine but the iGPU isn't cutting it anymore. Easy to add a discrete GPU if you have somewhere to put it.

The time has come to replace your machine. Now you have to transfer your 10TB of junk once. You don't need 100Gbps ethernet 99% of the time, but using the builtin gigabit ethernet for this is more than 24 hours of waiting. A pair of 100Gbps cards cuts that >24 hours down to ~15 minutes. If the old and new machines have an empty slot.

reply
My motherboard has 3 16x PCIe slots, but realistically only one is used for the GPU as the other two are under the mastodon of a cooler needed by the GPU. Can't use a 100G network card if I can't fit it under the GPU. Can't not use the GPU as I don't have an iGPU in my CPU.

He's not advocating from removing PCIe slots, but in practice, it's needed by way less consumers than before. There's probably more computers being sold right now without any PCIe slot than there are with more than 1.

reply
> My motherboard has 3 16x PCIe slots, but realistically only one is used for the GPU as the other two are under the mastodon of a cooler needed by the GPU.

Discrete GPUs generally consume two PCI slots, not three, and even the mATX form factor allows for four PCI slots (ATX is seven), which gives board makers an obvious thing to do. Put one x16 slot at the top and the other(s) lower down and use the space immediately under the top x16 slot for an x1 slot which is less inconvenient to block or M.2 slot which can be used even if there is a GPU hanging over it. This configuration is currently very common.

It also makes sense to put one of the x16 slots at the very bottom because it can either be used for a fast single height card (>1Gb network or storage controller) or a GPU in a chassis with space below the board (e.g. mATX board in ATX chassis) without blocking another slot.

reply
My post Mortem sentiments exactly. The lack of Nvidia GPU support for the M series Mac Pro models kneecapped the platform for professionals. If Apple had included that in those they’d be the defacto professional workstation for many more folks working in AI tech.
reply
On the other hand it forced developers to invest more in Metal which looks like an investment starting to bear fruit.
reply
Plus modern interconnects like CXL are also layers on top of PCIe, and USB4 supports PCIe tunnelling. PCIe is a big collection of specifications, the physical/link/transaction layers can be mixed and matched and evolved separately.

I don't see it disappearing, at most we'll get PCIe 6/7/etc.

reply
Thunderbolt is PCIe running over a cable.
reply
Sure, with expensive line drivers to send the data 1+ meters, instead of 10ish cm. And with only 2 channels instead of up to 16.
reply
Yes, I know; this is part of what I was implying when I said "Just about every single consumer computer shipped today uses PCIe."

I don't understand how this is a response to anything I said.

reply
Yup the 4090 and SoundBlaster ZXR in my AM5 7800X3D system would both like to upvote your reply.
reply
Sound card works fine on USB2 (RME for example has cards on USB2 that can manage 30/30 io at 192khz without issue at low latency if you have the CPU to deal with the load)

With USB3 you have 94 i/o…

For years pci has not been mandatory for audio. UAD, Apogee, RME and other high end brands will push you to them. Or even only provide them as usb device… even Thunderbolt is not needed here.

And that’s been the case for a while! My Fireface UC from 15 years ago can deal with 16 channels at 96khz at 256 sample. On PC and Mac.

reply
Personally, I'd love to see / read / hear more about the way RME do what they do. I know they basically update the fpga on the devices in lock step with the drivers, which allows them to do all sorts of magic (low CPU usage, zero latency recording of each raw channel being one of them) but I'd love an interview or article from some of the hardware and software people from RME. They have been rock solid and basically future proof for decades and I think the entire hardware and software industries could learn something from the way they do things.

Incredible products, definitely worth the premium.

reply
Then they should start putting internal high powered USB ports inside the case where I can literally bolt this shit into place because my desk is a goddamn mess of cables and dongles and boxes that don't stack or interlock or interface at all and I am so so utterly tired of being gaslit into beliving that they're just as good as a fucking slot.
reply
Sounds like a 9.5" mini rack could help with the stacking, see Geerling.
reply
I have about 14 or 15 USB devices in addition to my 4 monitors, and whilst I'm sure you're right I'm very happy to have a high quality soundcard that is not part of that mix.
reply
Compared to video data and the speed the CPU is running at, audio trickles in at a snails pace.
reply