Did you mean instead "The articles are okay if overly wordy"?
It's why I started to pay attention to what he says.
Dude is a bit verbose, but his rationale is solid. If it gets the panties of some people here in a bunch, he may be on to something.
Hot take, but really it's more of an observation than a take: We saw this exact response in Blockchain & crypto circles a few years ago. (Though HN wasn't quite as culturally "central" to those)
Economic Bubbles are subject to the Tinkerbell Effect. They exist so long as people exist in them, and collapse when either 1) They become so financially unsustainable as to collapse, having consumed all the money the economy could possibly give them, or 2) People stop believing in the bubble and stop feeding it money.
In this regard, the statement "NTFs are stupid" was not merely ridiculing those who bought them, but a direct attack on the bubble and those invested in it. And this is something the people involved in the bubble understand instinctively, even if they aren't consciously aware of it. (There's a psychological mechanism to that, but it's not relevant)
So consequently, they react aggressively to dissent. They seek to enforce their narrative, because not doing so is a threat to the bubble and their financial interests.
---
AI's not much different to that. It's clearly a bubble to everyone including the AI execs saying it out loud.
And people react aggressively to dissent like Ed's, because if the wider public stops believing in AI's future, the bubble bursts. They'll stop tolerating datacenter construction, they'll sell their Nvidia shares, they'll demand regulators restrict AI.
(And to those who can feel their aggression rising reading this comment. Hi, yes. I see you. If I were wrong, nothing I said would matter. You'd be wasting your time engaging with it, history would simply prove me wrong. But by all means, type up that reply or click that button.)
[I'm an AI-doomer myself, but I am an AI-doomer because by and large this stuff increasingly works, not because it doesn't.]
That said, Ed Zitron still does a lot of useful research into the economics of the industry and I also believe that continued progress in AI can disrupt the world (for better and for worse) while the economics propping up all the frontier model providers can also implode spectacularly.
Some people talk about how AI doom comes about either way because it could take all of our jobs OR crash the economy when the current bubble bursts. But as an uber-AI-doomer I happen to think there is a very real possibility of a double downside (for the labor class, at least) where both of those things can happen at the same time!