Yesterday "I don't know about you, but I benefit so much from using Claude at work that I would gladly pay $1,500-$2,000 per month to keep using it."
Thanks for that, and it's worth nothing FYI.
LLMs are probably the most impressive machine made in recorded human existence. Will there be a better machine? I'm 100% confident there will be, but this is without a doubt extremely valuable for a wide array of fields, including software development. Anyone claiming otherwise is just pretending at this point, maybe out of fear and/or hope, but it's a distorted view of reality.
By this do you mean there isn't much more room for future improvement, or that you feel it is not useful in its current form for software development? I think the latter is hard position to defend, speaking as a user of it. I am definitely more productive with it now, although I'm not sure I enjoy software development as much anymore (but that is a different topic)
I don't expect that LLM technology will improve in a way that makes it significantly better . I think the training pool is poisoned, and I suspect that the large AI labs have been cooking the benchmark data for years to suspect that their models are improving more quickly than they are in reality.
That being said, I'm sure some company will figure out new strategies for deploying LLMs that will cause a significant improvement.
But I don't expect that improvements are going to come from increased training.
> [Do] you feel it is not useful in its current form for software development?
IME using LLMs for software development corrodes my intuitive understanding of an enterprise codebase.
Since the advent of LLMs, I've been asked to review many sloppy 500+/1000+ line spam PRs written by arrogant Kool-Aid drinking coworkers. If someone is convinced that Claude Code is AGI, they won't hesitate to drop a slop bomb on you.
Basically I feel that coding using LLMs degrades my understanding of what I'm working on and enables coworkers to dominate my day with spam code review requests.
I feel you there, I definitely notice that. I find I can output high quality software with it (if I control the design and planning, and iterate), but I lack that intuitive feel I get about how it all works in practice. Especially noticeable when debugging; I have fewer "Oh! I bet I know what is going on!" eureka moments.
It seems so gross.
But I guess with all of the trillions of investor dollars being dumped into the businesses, it would be irresponsible to not run guerrilla PR campaigns
I think this takes away from the main thrust of your argument which is the marketing campaign and to me makes you seem conspiratorial minded. LLMs can be both useful and also mass astroturfing can be happening.
Personally I have witnessed non coders (people who can code a little but have not done any professional software building) like my spouse do some pretty amazing things. So I don’t think it’s useless.
It can be all of:
1. It’s useful for coding
2. There’s mass social media astroturfing happening
3. There’s a massive social overhype train that should be viewed skeptically
4. Theres some genuine word of mouth and developer demand to try the latest models out of curiosity, with some driven by the hype train and irrational exuberance and some by fear for their livelihoods.
IN MY GENUINELY HELD OPINION, LLMs generate shit code and the people who disagree don't know what good code looks like.
Yes they are. This is true.
> which is a time consuming and tedious part of programming.
In my experience, this is a tedious part of programming which I do not spend very much time on.
In my experience LLM generated API boilerplate is acceptable, yet still sloppier than anything I would write by hand.
In my experience LLMs are quite bad at generating essentially every other type of code, especially if you are not generating JS/TS or HTML/CSS.