upvote
Wow, just like its larger brother!

I had a similarly bad experience running Qwen 3.5 35b a3b directly through llama.cpp. It would massively overthink every request. Somehow in OpenCode it just worked.

I think it comes down to temperature and such (see daniel‘s post), but I haven’t messed with it enough to be sure.

reply
You're not doing anything wrong, that's expected
reply