Two years ago, one former exec at my place was perfectly happy to throw resources ( his word ) from India at a problem, while unwilling to pay the vendor for the same thing. I voiced my objection once, but after it was dismissed I just watched the thing blow up.
I am not saying current situation is the same. It is not. But, it is the same hubris, which means miscalculations will happen ( like with Dorsey's Block mass firing ).
Now I've come to realize the error in my ways, this is probably not going to happen. What will happen is instead is that the ones doing the "shuffling of shit" is just going to also be agents themselves. Prompted by a more senior slop-grammer specialized in orchestrating "shuffling of shit".
This task was famously incredibly difficult back when we had people producing unmaintainable mountains of millions of lines of code, to the point where shipping anything sizable in a working state on time without last minute scope reductions is nearly unheard of.
I can't imagine using AI to add another one to two zeroes to the lines of code counter would help reach the goal post.
LLMs can write a lot of code. they can even write a comprehensive test suite for that code. However they can't tell you if it doesn't work because of some interaction with something else you didn't think about. They can't tell you that all race conditions are really fixed (despite being somewhat good at tracking them down when known). They can't tell you that the program doesn't work because it doesn't do something critical that nobody thought to write into the requirements until you noticed it was missing.