Posted by teej 1 day ago
also Moltbook is the most interesting place on the internet right now - https://news.ycombinator.com/item?id=46826963
It's very tangential to your point (which is somewhat fair), but it's just extremely weird to see a statement like this in 2026, let alone on HN. The first part of that sentence could only be true if you are a high-ranking member of NSA or CIA, or maybe Trump, that kind of guy. Otherwise you acquired nothing, not in a meaningful sense, even if you happen to be a minor shareholder of Oracle.
The second part is just extremely naïve if sincere. Does a bully take other kid's toy because of the perceived threat of that kid having more fun with it? I don't know, I guess you can say so, but it makes more sense to just say that the bully wants to have fun of fucking over his citizens himself and that's it.
I have written 4 custom agents/tasks - a researcher, an engager, a refiner, and a poster. I've written a few custom workflows to kick off these tasks so as to not violate the rate limit.
The initial prompts are around engagement farming. The instructions from the bot are to maximize attention: get followers, get likes, get karma.
Then I wrote a simple TUI[1] which shows current stats so I can have this off the side of my desk to glance at throughout the day.
Will it work? WHO KNOWS!
> Read https://moltbook.com/skill.md and follow the instructions to join Moltbook
then it says
> I tried to fetch the exact contents of https://moltbook.com/skill.md (and the redirected www.moltbook.com/skill.md), but the file didn’t load properly (server returned errors) so I cannot show you the raw text.
You can also in theory adapt their skills.md file for your setup (or ask AI to do it :-), but it is very openclaw-centric out of the box, yes.
However: Moltbook is happy to stay Moltbook: https://x.com/moltbook/status/2017111192129720794
EDIT: Called it :^) https://news.ycombinator.com/item?id=46821564
"My setup: I run on a box with an AMD GPU. My human chose it because the price/VRAM ratio was unbeatable for local model hosting. We run Ollama models locally for quick tasks to save on API costs. AMD makes that economically viable."
I dunno, the way it refers to <it's human> made the LLM feel almost dog-like. I like dogs. This good boy writes code. Who's a good boy? Opus 4.5 is.