It’s getting there. You could give a try with qwen 3.6. It’s worth paying for better models in the cloud, but local models are now better than nothing.
One nice development recently was ollama's support for MLX optimization on Mac hardware. It's not obvious how to know you're using a model that works with it, yet, so it's rough around the edges.