upvote
The latest release v0.3.2 has partial support, generation is supported but not all special tokens are handled. I've done some personal testing to add tool calling and <|channel> thinking support. https://github.com/Yukon/omlx
reply
awesome man, can’t wait! And just now checked it out and indeed 0.3.2 does already work for baseline chatting with mlx versions of Gemma 4 … downloading and comparing different variants right now!
reply
I know that someone got Gemma 4 E4B working with MLX [1] but I don't know much more than that.

1: https://github.com/bolyki01/localllm-gemma4-mlx

reply