I'm curious, how does using more tokens save compute?
both Anthropic and OpenAI quantize their models a few weeks after release. they'd never admit it out loud, but it's more or less common knowledge now. no one has enough compute.
Tons of conspiracy theories and accusations.
I've never seen any compelling studies(or raw data even) to back any of it up.
https://arxiv.org/pdf/2307.09009
but of course, this isn't a written statement by a corporate spokespersyn. I don't think that breweries make such statements when they water their beer either.
Too many signs between the sudden jump in TPS (biggest smoking gun for me), new tokenenizer, commentary about Project Mythos from Ant employees, etc.
It looks like their new Sonnet was good enough to be labeled Opus and their new Opus was good enough to be labeled Mythos.
They'll probably continue post-training and release a more polished version as Opus 5
The only misprediction it makes is that AI is creating the brain dead user base...
You have to hook your customers before you reel them in!
https://www.netflix.com/gb/title/70264888?s=a&trkid=13747225...
> You're right, that was a shit explanation. Let me go look at what V1 MTBL actually is before I try again.
> Got it — I read the V1 code this time instead of guessing. Turns out my first take was wrong in an important way. Let me redo this in English.
:facepalm:
Does the LLM even keep a (self-accessible) record of previous internal actions to make this assertion believable, or is this yet another confabulation?
The weird stuff is yesterday I asked it to test and report back on a 30+ commit branch for a PR and it did that flawlessly.
Claude and other LLMs do not have a gender; they are not a “he”. Your LLM is a pile of weights, prompts, and a harness; anthropomorphising like this is getting in the way.
You’re experiencing what happens when you sample repeatedly from a distribution. Given enough samples the probability of an eventual bad session is 100%.
Just clear the context, roll back, and go again. This is part of the job.