upvote
I’ve been using 1M for a while and it defers it and makes it worse almost when it happens. Compacting a context that big loses a ton of fidelity. But I’ve taken to just editing the context instead (double esc). I also am planning to build an agent to slice the session logs up into contextually useful and useless discarding the useless and keeping things high fidelity that way. (I.e., carve up with a script the jsonl and have subagent haiku return the relevant parts and reconstructing the jsonl)
reply
til you can edit context. i keep a running log and /clear /reload log
reply
double escape gets you to a rewind. not sure about much else.

the conversation history is a linked list, so you can screw with it, with some care.

I spend this afternoon building an MCP do break the conversation up into topics, then suggest some that aren't useful but are taking up a bunch of context to remove (eg iterations through build/edit just needs the end result)

its gonna take a while before I'm confident its worth sharing

reply
Yeah just selective rewind. Selective edit where you elide large token sinks of coding and banging its head on the wall is what u mean. Not something I’ve seen done yet but there’s no reason - I suspect if you do a token use distribution in programming session most goes to pretty low semantic value malarkey.
reply
yea i thought session managment was some sort of secret sauce.

I keep a running log of important things and then i just clear context and reload that file into context.

would that work

reply