upvote
Totally agree, I mentioned it in another comment but Gemini was a game changer for allowing me to increase the size of the project I can feasibly have AI work on.

Only issue is Gemini's context window (I've seen my experience corroborated here on HN a couple times) isn't consistent. Maybe if 900k tokens are all of unique information, then it will be useful to 1 million, but I find if my prompt has 150k tokens of context or 50k, after 200k in the total context window response coherence and focus goes out the window.

reply
I'd love some more innovation on increasing context size without blowing up RAM usage. Mistral small 2503 24B and Gemma 3 27B both fit into 24GB at Q4, but Mistral can only go up to about 32k and Gemma about 12k before all VRAM is exhausted, even with flash attention and KV cache quantization.
reply
What editor are you using with gemini 2.5 pro? I really don't like their vscode extension.
reply