upvote
It’s getting there. You could give a try with qwen 3.6. It’s worth paying for better models in the cloud, but local models are now better than nothing.
reply
One nice development recently was ollama's support for MLX optimization on Mac hardware. It's not obvious how to know you're using a model that works with it, yet, so it's rough around the edges.

https://ollama.com/blog/mlx

reply
Use llama.cpp or better yet Unsloth Studio
reply