there’s some debate about whether this is in the spirit of the _original_ Ralph, because it keeps too much context history around. But in practice Claude Code compactions are so low-quality that it’s basically the same as clearing the history every few turns
I’ve had good luck giving it goals like “keep working until the integration test passes on GitHub CI” - that was my longest run, actually, it ran unattended for 24 hours before solving the bug
I still have a manual part which is breaking the design document down into multiple small gh issues after a review but I think that is fine for now.
Using codex exec, we start working on a github issue with a supplied design document, creating a PR on completion. Then we perform a review using a review skill madeup which is effectively just a "cite your sources" skill on the review along with Open Questions.
Then we iterate through open questions doing a minimum of 3 reviews (somewhat arbitrary but sometimes multiple reviews catch things). Then finally I have I have a step in for checking Sonarcloud, fixing them and pushing the changes. Realistically this step should be broken out into multiple iterations to avoid large context rot.
What I miss the most is output, seeing whats going on in either Codex or Claude in real time. I can output the last response but it just gets messy until I make something a bit more formal.
The key bit is right under that though. Ralph is literally just this:
while :; do cat PROMPT.md | npx --yes @sourcegraph/amp ; done cat PROMPT.md | cat | npx --yes @sourcegraph/ampAnd it all ends with the grift of all grifts: promoting a crypto token in a nonchalant 'hey whats this??!!??' way...
It's complete garbage, and since it runs in a loop, the amount of garbage multiplies over time.
Even if in the absolute the ceiling remains low, it’s interesting the degree to which good context engineering raises it
I’m all for AI and it’s evident that the future of AI is more transparency (MLOPs, tracing, mech interp, AI safety) not less.
And this dream of "having Claude implement an entire project from start to finish without intervention" came crashing down with this realization: Coding assistants 100% need human guidance.
More generally, I've noticed that people who spend a lot of time interacting with LLMs sometimes develop a distinct brain-fried tone when they write or talk.