As an everyday user of AI, both at work and privately, I am not that convinced. The biggest effect I've seen so far is demand for faster work because "everything is faster with agents", but software quality is slowly dropping in software I see around me.
Current AI is very useful as a trivia engine and as a language manipulation tool - i.e. it can quickly extract information from a huge amount of text. But it still sucks when writing new things.
Admittedly, here has been much progress, but it seems to be slowing down. Money is drying out, models are getting nerfed, and only better scaffolding and workflows are making it better. Unless they build 100x more data centers, I don't see models getting significantly better.
Yes? Legitimately curious what other explanation is there here, thats the reason all of these LLM integrations across all software is being pushed.
Like this.[1]
> AI-assisted coding has become the norm and with tools like Cursor, GitHub Copilot, Claude Code, Codex, we are increasingly letting models touch our code.
... how is it good literaly style to both (1) assume that something is the norm, and (2) use a long intro-sentence to state that something is the norm? Pick a lane—either it is the norm and you don’t need to state it or it isn’t and you need to set the stage. Stating the apparently obvious makes your (their) writing read like a eighth grade paper.
In short I’m agnostic as far as proclamations go. ;)