upvote
yeah this is mindblowing speed. imagine this with opus 4.6 or gpt 5.2. probably coming soon
reply
I'd be happy if they can run GLM 5 like that. It's amazing at coding.
reply
Why do you assume this?

I can produce total jibberish even faster, doesn’t mean I produce Einstein level thought if I slow down

reply
Better models already exist, this is just proving you can dramatically increase inference speeds / reduce inference costs.

It isn't about model capability - it's about inference hardware. Same smarts, faster.

reply
Not what he said.
reply