upvote
Composer-2 is based on Kimi K2.5, but with extensive RL. Cursor estimated 3x more compute on their RL than the original K2.5 training run (some details in https://cursor.com/blog/composer-2-technical-report).

Composer-2 seems very useful in Cursor, while K2.6 according to AA seems to be a really useful general model: https://artificialanalysis.ai/articles/kimi-k2-6-the-new-lea...

reply
I'm going to be brutally honest but I have not found Kimi to be useful at all. It simply cannot compete with what closed models from Codex and Claude offers. I don't want to risk using a model outside the ecosystem and introduce variables as most of my workflow is baked into two to three large company models.
reply
That's interesting, Kimi K2.5 used through KimiCode was comparable to Sonnet in my tests, and is an excellent alternative to Anthropic models

That being said, I noticed that Kimi being served through Openrouter providers was trash. Whatever they do on the backend to optimize for throughput really compromised the intelligence of the model. You have to work with Kimi directly if you want the best results, and that's also probably why they released a test suite to verify the intelligence of their new models.

reply
Kimi is my favorite of the Chinese models.

I found it much more consistent than glm or minimax

reply
Which version of Kimi and served from where?
reply
On the other hand, I found MiniMax M2.7 a reasonable model that I could trust.

I guess really depends on tastes

reply
deleted
reply