I'm curious, how does using more tokens save compute?
both Anthropic and OpenAI quantize their models a few weeks after release. they'd never admit it out loud, but it's more or less common knowledge now. no one has enough compute.
Tons of conspiracy theories and accusations.
I've never seen any compelling studies(or raw data even) to back any of it up.
https://arxiv.org/pdf/2307.09009
but of course, this isn't a written statement by a corporate spokespersyn. I don't think that breweries make such statements when they water their beer either.
Too many signs between the sudden jump in TPS (biggest smoking gun for me), new tokenenizer, commentary about Project Mythos from Ant employees, etc.
It looks like their new Sonnet was good enough to be labeled Opus and their new Opus was good enough to be labeled Mythos.
They'll probably continue post-training and release a more polished version as Opus 5
The only misprediction it makes is that AI is creating the brain dead user base...
You have to hook your customers before you reel them in!
https://www.netflix.com/gb/title/70264888?s=a&trkid=13747225...