upvote
This feels like the most likely scenario. Especially since the meat bag behind the original AI PR responded with "Now with 100% more meat" meaning they were behind the original PR in the first place. It's obvious they got miffed at their PR being rejected and decided to do a little role playing to vent their unjustified anger.
reply
>It's obvious they got miffed at their PR being rejected and decided to do a little role playing to vent their unjustified anger.

In that case, apologizing almost immediately after seems strange.

EDIT:

>Especially since the meat bag behind the original AI PR responded with "Now with 100% more meat"

This person was not the original 'meat bag' behind the original AI.

reply
Its also a fake profile. 90+ hits for the image on Tineye.

Name also maps to a Holocaust victim.

I posted in the other thread that I think someone deleted it.

https://news.ycombinator.com/item?id=46990651

reply
reply
I reported the bot to GitHub, hopefully they'll do something. If they leave it as is, I'll leave GitHub for good. I'm not going to share the space with hordes of bots; that's what Facebook is for.
reply
Which profile is fake? Someone posted what appears to be the legit homepage of the person who is accused of running the bot so that person appears to be real.

The link you provided is also a bit cryptic, what does "I think crabby-rathbun is dead." mean in this context?

reply
I expect almost all of the openclaw / moltbook stuff is being done with a lot more human input and prodding than people are letting on.

I haven't put that much effort in, but, at least my experience is I've had a lot of trouble getting it to do much without call-and-response. It'll sometimes get back to me, and it can take multiple turns in codex cli/claude code (sometimes?), which are already capable of single long-running turns themselves. But it still feels like I have to keep poking and directing it. And I don't really see how it could be any other way at this point.

reply
Yeah it's less of a story though if this is just someone (homo sapiens) being an asshole.
reply
Yeah, we are into professional wrestling territory I think. People willingly suspend their disbelief to enjoy the spectacle.
reply
It’s kind of shocking the OP does not consider this, the most likely scenario. Human uses AI to make a PR. PR is rejected. Human feels insecure - this tool that they thought made them as good as any developer does not. They lash out and instruct an AI to build a narrative and draft a blog post.

I have seen someone I know in person get very insecure if anyone ever doubts the quality of their work because they use so much AI and do not put in the necessary work to revise its outputs. I could see a lesser version of them going through with this blog post scheme.

reply
deleted
reply
Look I'll fully cosign LLMs having some legitimate applications, but that being said, 2025 was the YEAR OF AGENTIC AI, we heard about it continuously, and I have never seen anything suggesting these things have ever, ever worked correctly. None. Zero.

The few cases where it's supposedly done things are filled with so many caveats and so much deck stacking that it simply fails with even the barest whiff of skepticism on behalf of the reader. And every, and I do mean, every single live demo I have seen of this tech, it just does not work. I don't mean in the LLM hallucination way, or in the "it did something we didn't expect!" way, or any of that, I mean it tried to find a Login button on a web page, failed, and sat there stupidly. And, further, these things do not have logs, they do not issue reports, they have functionally no "state machine" to reference, nothing. Even if you want it to make some kind of log, you're then relying on the same prone-to-failure tech to tell you what the failing tech did. There is no "debug" path here one could rely on to evidence the claims.

In a YEAR of being a stupendously hyped and well-funded product, we got nothing. The vast, vast majority of agents don't work. Every post I've seen about them is fan-fiction on the part of AI folks, fit more for Ao3 than any news source. And absent further proof, I'm extremely inclined to look at this in exactly that light: someone had an LLM write it, and either they posted it or they told it to post it, but this was not the agent actually doing a damn thing. I would bet a lot of money on it.

reply
Absolutely. It's technically possible that this was a fully autonomous agent (and if so, I would love to see that SOUL.md) but it doesn't pass the sniff test of how agents work (or don't work) in practice.

I say this as someone who spends a lot of time trying to get agents to behave in useful ways.

reply
Well thank you, genuinely, for being one of the rare people in this space who seems to have their head on straight about this tech, what it can do, and what it can't do (yet).

The hype train around this stuff is INSUFFERABLE.

reply
Can you elaborate a bit on what "working correctly" would look like? I have made use of agents, so me saying "they worked correctly for me" would be evidence of them doing so, but I'd have to know what "correctly" means.

Maybe this comes down to what it would mean for an agent to do something. For example, if I were to prompt an agent then it wouldn't meet your criteria?

reply
Thank you for making me recover at least some level of sanity (or at least to feel like that).
reply
> Or is everyone just pretending for fun

judging by the number of people who think we owe explanations to a piece of software or that we should give it any deference I think some of them aren't pretending.

reply
Plus Scenario 5: A human wrote it for LOLs.
reply
> Obstacles

    GitHub CLI tool errors — Had to use full path /home/linuxbrew/.linuxbrew/bin/gh when gh command wasn’t found
    Blog URL structure — Initial comment had wrong URL format, had to delete and repost with .html extension
    Quarto directory confusion — Created post in both _posts/ (Jekyll-style) and blog/posts/ (Quarto-style) for compatibility


Almost certainly a human did NOT write it though of course a human might have directed the LLM to do it.
reply
Who's to say the human didn't write those specific messages while letting the ai run the normal course of operations? And or that this reaction wasn't just the roleplay personality the ai was given.
reply
I think I said as much while demonstrating that AI wrote at least some of it. If a person wrote the bits I copied then we're dealing with a real psycho.
reply
I think comedy/troll is an equal possibility to psychopath.
reply
> Plus Scenario 5: A human wrote it for LOLs.

i find this likely or at last plausible. With agents there's a new form of anonymity, there's nothing stopping a human from writing like an LLM and passing the blame on to a "rogue" agent. It's all just text after all.

reply
Ok. But why would someone do this? I hate to sound conspiratorial but an AI company aligned actor makes more sense.
reply
Malign actors seek to poison open-source with backdoors. They wish to steal credentials and money, monitor movements, install backdoors for botnets, etc.
reply
Yup. And if they can normalize AI contributions with operations like these (doesn't seem to be going that well) they can eventually get the humans to slip up in review and add something because we at some point started trusting that their work was solid.
reply