upvote
The idea is that smarter models might use fewer turns to accomplish the same task - reducing the overall token usage

Though, from my limited testing, the new model is far more token hungry overall

reply
Well you‘ll need the same prompt for input tokens?
reply
Only the first one. Ideally now there is no second prompt.
reply
Are you aware that every tool call produces output which also counts as input to the LLM?
reply
deleted
reply
That's valid, but it's also worth knowing it's only one part of the puzzle. The submission title doesn't say "input".
reply