upvote
I failed to run in LM Studio on M5 with 32gb at even half max context. Literally locked up computer and had to reboot.

Ran gemma-4-26B-A4B-it-GGUF:Q4_K_M just fine with llama.cpp though. First time in a long time that I have been impressed by a local model. Both speed (~38t/s) and quality are very nice.

reply
Tool calls falling is a problem with the inference engine’s implementation and/or the quant. Update and try again in a few days.

This is how all open weight model launches go.

reply
Haven't had time to try yet, but heard from others that they needed to update both the main and runtime versions for things to work.
reply
Even with the latest version of LM Studio and the latest runtimes I find that tool use fails 100% of the time with the following error: Error rendering prompt with jinja template: "Cannot apply filter "upper" to type: UndefinedValue".

EDIT: The issue is addressed in LM Studio 0.4.9 (build 1), which auto-update wasn't picking up for me for some reason.

reply
reply
Alas, this does not resolve the issue for me.
reply