upvote
> It also runs on my own computer, and the latest frontier open source models are able to drive it (Kimi, etc). The future is going to be locally hosted and ad free and there’s nothing Big Tech can do about it. It’s glorious.

After messing with openclaw on an old 2018 Windows laptop running WSL2 that I was about to recycle, I am coming to the same conclusion, and the paradigm shift is blowing my mind. Tinkerers paradise.

The future is glorious indeed.

reply
Same here. I like tinkering with my Home Assistant setup and small web server running miscellaneous projects on my Raspberry Pi, but I hate having to debug it from my phone when it all falls over while I'm not near my computer.

Being able to chat with somebody that has a working understanding of a Unix environment and can execute tasks like "figure out why Caddy is crash looping and propose solutions" for a few dollars per month is a dream come true.

I'm not actually using OpenClaw for that just yet, though; something about exposing my full Unix environment to OpenAI or Anthropic just seems wrong, both in terms of privacy and dependency. The former could probably be solved with some redacting and permission-enforcing filter between the agent and the OS, but the latter needs powerful local models. (I'll only allow my Unix devops skills to start getting rusty once I can run an Opus 4.5 equivalent agent on sub-$5000 hardware :)

reply
What if you don't want to tinker? You just want something that works. Is it still transformative?
reply
Honest answer: OpenClaw still requires some tinkering, but it's getting easier.

The biggest barrier for non-tinkerers is the initial setup - Node.js, API keys, permissions, etc. Once it's running, day-to-day use is pretty straightforward (chat with it like any other messaging app).

That said, you'll still need to: - Understand basic API costs to avoid surprises - Know when to restart if it gets stuck - Tweak settings for your specific use case

If you're determined to skip tinkering entirely, I'd suggest starting with just the messaging integration (WhatsApp/Telegram) and keeping skills/tools minimal. That's the lowest-friction path.

For setup guidance without deep technical knowledge, I found howtoopenclawfordummies.com helpful - it's aimed at beginners and covers the common gotchas.

Is it transformative without tinkering? Not yet. The magic comes from customization. But the baseline experience (AI assistant via text) is still useful.

reply
> This is the product that Apple and Google were unable to build

It's not they're unable to build it, it's that their businesses are built on "engagement" and wasting human time. A bot "engaging" with the ads and wasting its time would signal the end of their business model.

reply
Yes, they already said that.

> because it’s a threat to their business model.

reply
What model, or models, are you querying in the backend?
reply
How are you running Kimi locally?
reply
Quantized, heavily, and offloading everything possible to sysram. You can run it this way, just barely reachable with consumer hardware with 16 to 24gb vram and 256gb sysram. Before the spike in prices you could just about build such a system for $2500, but the ram along probably adds another $2k onto that now. Nvidia dgx boxes and similar setups with 256gb unified ram can probably manage it more slowly ~1-2 tokens per second. Unsloth has the quantized models. I’ve test Kimi though don’t have quite the headroom at home for it, and I don’t yet see a significant enough difference between it and the Qwen 3 models that can run in more modest setups: I get a highly usable 50 tokens per second out of the A3B instruct that fits into 16gb VRAM with enough left over not to choke Netflix and other browser tasks, it performs on par with what I ask out of Haiku in Claude Code, and better as my own tweaking improves with the also ever better tooling that comes out near weekly.
reply
I have an AMD Epyc machine with 512GB of RAM and a humble NVIDIA 3090. You will have to run a quantized version but you can get a couple tokens per second out of it since these models are optimized to split across the GPU/RAM and it's about as good as Claude was 12 months ago.

Full disclosure, I use OpenRouter and pay for models most of the time since it's more practical than 5-10 tokens per second, but the option to run it "If I had to, worst case" is good enough for me. We're also in a rapidly developing technology space and the models are getting smaller and better by the day, ever year the smaller models get better

reply
> The future is going to be locally hosted and ad free and there’s nothing Big Tech can do about it.

I wouldn't be so certain of that. Someone is paying to train and create these models. Ultimately, the money to do that is going to have to come from somewhere.

reply
Good News! The Models are done, and you can download them for free. Even if they stopped being worked on this moment, those are finished and usable right now, and won't get any worse over time.
reply
They wouldn't get any worse, but I assume they'd get behind really fast.
reply
Old models become outdated. Not sure how useful an outdated model can still be.
reply
They can still be very useful because new models are reaching an asymptote in performance. Meanwhile as hardware gets cheaper in the future (current RAM prices notwithstanding), these models will become faster to run on local hardware.
reply
P2P training is a possibility. With so much compute running the block chain I can see this utilized for AI.

Big tech is bought and paid for by consumers. We can do the same for oss trained models.

reply
I don’t get it. You can do that with the Claude app or ChatGPT too. What’s the value add?

Edit: oh I see. It’s local. So privacy. Quite a good value add actually.

reply
What privacy? If you're using ChatGPT or Claude, your chats are still logged.
reply
It's local, meaning it uses local models, what they said in the sentence prior to the privacy one.
reply
Unless you have unusually powerful hardware, local models will unfortunately currently not really cut it for Moltbot.
reply
OP implied they have powerful enough hardware, since Kimi runs on their computer, so that is why they mentioned it is local. That it doesn't work for most people has no relation to what OP of this thread said. Regardless, you don't need an Opus level model, you can use a smaller one that'll just be slower at getting back to you, it's all asynchronous anyway compared to a coding agent where some level of synchronicity is expected.
reply
GGP seem to be under the misapprehension that privacy is a core aspect/advantage of OpenClaw, when for most users it's really not.

So yes, I think the majority user experience is very relevant.

reply
This thread is about giving one's opinions on their personal experiences with the tool, so the OP of the thread can say whatever they want, it doesn't mean they think it's at all related to the "majority user experience" nor do they have to cater their opinions towards that.
reply
GLM 4.7-flash does very well although OpenClaw has some work to do for CoT.
reply
From what I've read, OpenClaw only truly works well with Opus 4.5.
reply
The latest Kimi model is comparable in performance at least for these sorts of use cases, but yes it is harder to use locally.
reply
> harder to use locally

Which means most people must be using OpenClaw connected to Claude or ChatGPT.

reply
It's the other way around. At least for most people, it grants access to your personal data to an LLM (and by extension its inference provider) in the cloud.
reply