upvote
Yeah this seems to be a very bad idea. Seems like the author had the right idea, but the wrong way of implementing it.

There are a few papers actually that describe how to get faster results and more economic sessions by instructing the LLM how to compress its thinking (“CCoT” is a paper that I remember, compressed chain of thought). It basically tells the model to think like “a -> b”. There’s loss in quality, though, but not too much.

https://arxiv.org/abs/2412.13171

reply
For the more important sessions, I like to have it revise the plan with a generic prompt (e.g. "perform a sanity check") just so that it can take another pass on the beginning portion of the plan with the benefit of additional context that it had reasoned out by the end of the first draft.
reply
Is this true? Non-reasoning LLMs are autoregressive. Reasoning LLMs can emit thousands of reasoning tokens before "line 1" where they write the answer.
reply
reasoning is just more tokens that come out first wrapped in <thinking></thinking>
reply
They are all autoregressive. They have just been trained to emit thinking tokens like any other tokens.
reply
there are no reasoning LLMs.
reply
This is an interesting denial of reality.
reply
A "reasoning" LLM is just an LLM that's been instructed or trained to start every response with some text wrapped in <BEGIN_REASONING></END_REASONING> or similar. The UI may show or obscure this part. Then when the model decides to give its "real" response, it has all that reasoning text in its context window, helping it generate a better answer.
reply
deleted
reply
I don't think Claude Code offers no thinking as an option. I'm seeing "low" thinking as the minimum.
reply
Ugh. Dictated with such confidence. My god, I hate this LLMism the most. "Some directive. Always this, never that."
reply