upvote
A suggestion: Don't invest in any new hardware to run an LLM locally until you've tried the model for a while through OpenRouter.

The Qwen models are cool, but if you're coming from Opus you will be somewhere between mildly to very disappointed depending on the complexity of your work.

reply
OpenRouter-served models are often more heavily quantized than what you can run locally, or try for yourself on generic cloud-based infrastructure.
reply
Been having a ton of fun with copilot cli directed to local qwen 3.6. If you’re willing to increase the amount of specificity in your prompts then delegating from a GPT-5.4 or Opus to local qwen has been great so far.
reply