One interesting variable is that I'm located in Vietnam while my coworkers are located in Norway and Europe.
To work around this issue I used Claude for coding with a Copilot subscription which was much cheaper and had virtually no rate limiting.
Copilot gives you some set amount of credits each month, but you can also pay as you go if you run out of credit which is much better than the 5 hour window crap claude code would give me.
The only opus model available now on copilot for some reason is 4.7 and it costs 7.5x tokens, while everything else is 1x, 0.33x or free.
But I switched to using GPT 5.4 medium for a month or so which I find very reasonable.
At this rate I fully anticipate being able to run a comparable stack on a 128GB Mac Studio using quants of newer-generation distilled OSS models in a year or two. Being able to ramble to a computer for an hour about features and technical philosophy then have it build a nearly-working app for $50 is an exciting feeling. There's still a long tail of productionization and fixing what the model didn't adhere to but it's still incredible.
I got the 20$ gpt tier, and now i just use claude to craft MD plan docs instead, and then i hand them off to gpt 5.4 and it has been working great. can do about 4x as much work or so based on my feelings(not accurate). if i have just small simple stuff to do i might still fire those off with sonnet and that seems plenty viable, but as soon as its an opus tier task i swap to this workflow.
Little annoying as now im kinda trying to manage a .claude/ and an .opencode/ folder but i kinda just have the .opencode/ stuff reference the .claude/ stuff so its a little less bleh.
I've been keeping within my usage because ive been in a funk a bit, but when i was slightly more worried id sorta just juggle whether claude or gpt would handle writing some initial tests as it did seem to kinda be imbalanced otherwise. seems like gpt just spam resets weekly usage throughout the week anyway so its prolly nbd.
There is a lot of political capital to be earned by appearing to be "tough" on AI companies.
Glad I’m not the only one!
I’ve been limited so often this week I’ve setup half a dozen token compression tools in my workflow and had to do a crash course in token optimization.
Of course, it seems to only slightly delay the inevitable and doesn’t really solve the problem.