If you have an M processor then I would recommend that you ditch Ollama because it performs slowly. We get double or triple tok/s using omlx or vmlx, respectively, but vmlx doesn't have extensive support for some models like gpt-oss.
you can tell gemma4 comes from gemini-3
Opencode was able to create the library as well. It just took about 2x longer.
Next week I will be trying qwopus 27b.