upvote
> But writing a genuinely good harness with lots of context engineering and solid tool integration is in fact not that easy.

true, but its not worth $60 billion fucking quid.

reply
it's insanity.

the whole thing is driven by irrational stock market investers who NEED ai to be the thing that saves the world.

they're betting everything on it.

reply
deleted
reply
There are plenty of harder things in the world and very few are worth 60B.
reply
> (...) writing a genuinely good harness with lots of context engineering and solid tool integration is in fact not that easy.

This. They are after the harness engineering experience of the Cursor people, I'd assume the they want to absorb all that into Grok's offerings.

The value and the room for innovation on the harness side seems to be underestimated.

Oddly the harness also affects model training, since even GLM/Z.ai for example train (I suspect) their model on the actual Claude Code harness. So the choises made by harness engineers affects the model. For Kimi/Moonshot and OpenAI the company makes their own harness. Alibaba uses Gemini.

Very interesting dynamics.

reply
Isn't Codex TUI available for free though? Besides others like Pi and OpenCode of course.
reply
It can use local/oss models, but it doesn't make it simple to do (easiest with ollama) and it's not clear what else you 'lose' by making that choice.

If you had a really good (big) local model, maybe it's an option, but on the more common smaller (<32b) models, it will have similar problems in looping, losing context, etc. in my experience.

It's a nice TUI, but the ecosystem is what makes it good.

reply