Top
Best
New

Posted by brendanmc6 6 hours ago

Specsmaxxing – On overcoming AI psychosis, and why I write specs in YAML(acai.sh)
142 points | 136 commentspage 2
jwpapi 4 hours ago|
And once you’ve written all these specs you realize it became so slow that it’s faster to do it yourself in editor
girvo 3 hours ago||
People don’t actually track wall clock time, I’ve noticed.
motoxpro 4 hours ago|||
at which point you realize you never had a plan written down and you are using the code as a spec
jbjbjbjb 4 hours ago||
Which takes us back to this:

https://haskellforall.com/2026/03/a-sufficiently-detailed-sp...

wiseowise 3 hours ago||
But have you thought about “fun factor”? It’s where you sit like an addict in a casino for weeks and burn tokens in a hope of winning a software that you could’ve written? Who doesn’t consider “fun” thinking about work crap all the time, writing to your agent, verifying walls of slop?
teekert 2 hours ago||
It’s like a yaml of an event model but less graphical. Right? I think I will prefer Event Modeling especially with Martin Dilger now building tooling very much with agents in mind. There is no one place to read about his most recent efforts except for his LinkedIn feed though I fear.l so won’t post any urls, but information is easy enough to find.

A full blown event model facilitates all communication, human (management, devs, ops) and agentic. But maybe I’m missing something, maybe the dashboard can have this function I didn’t dig into it too much.

wesselbindt 4 hours ago||
I'm still confused as to why folks don't just write executable specs.
mike_hearn 2 hours ago||
Some of us do! That's called Gherkin.
eterm 4 hours ago|||
Ambiguity is the grease that keeps everything turning.
cenamus 4 hours ago|||
So basically tests?
MoreQARespect 4 hours ago||
Yes, except a test can be turing complete - i.e. code.

An executable spec like gherkin or hitchstory is config - it has no loops or conditionals. There are a number of rarely recognized benefits to this.

carlbarrdahl 4 hours ago|||
Could you expand on this?
booi 4 hours ago||
code
arikrahman 4 hours ago||
Literate programming would provide specs and code instead of working backwards from hard coded functions to figure out specs.
cassianoleal 2 hours ago||
> working backwards from hard coded functions to figure out specs.

People do that? Actual professionals?

fudgeonastick 4 hours ago||
If you're confused, and have tried Opus for coding, I'm keen to hear what problems or workflows it's not good at.

If you're genuinely confused, and haven't tried Opus for coding, then it's not surprising you're confused!

It is also okay for you to just not like the idea of LLMs for coding (but say that!).

wiseowise 3 hours ago|||
I’m using Opus 4.6 and I’m so confused! Maybe I should try Opus 4.7, which is almost twice as expensive to get some clarity (but not too much, I need to save money for Opus 4.8)?
oytis 4 hours ago|||
That's what the article is about - overcoming problems with AI cooding tools using specs in Yaml. If we've got that far, it might be better to write specs in a proper programming language instead and skip the AI layer altogether
adi_kurian 3 hours ago||
Think the idea is to still get monumental acceleration between fancy YAML specs (bullet points with some indentation that an intelligent technical manager could write) and production ready code.
taffydavid 3 hours ago||
I just spent a week training up in spec driven development through bmad, which was awful, and speckit which was ok but not great. Both had what seemed like unnecessary ceremony around the specs, generating fields of spec documents which presumably fill up the context window quickly. I just kept thinking "this should be using something simpler, all this markdown is unnecessary"

This seems like the answer to that thought!

k9294 2 hours ago||
What is yours agentic development experience with elixir? I used to like elixir a lot during a pre agentic era, but with coding agents it feels like the language isn't the best choice - slow compile time, weak type system (at least it was a year ago, I know there is work on that front), small ecosystem...
augment_me 4 hours ago||
Completely subjective take, but I feel like 95% of these "tools" that are prompt-engineering inventions created by the authors with their bias and to suit their needs don't have anything supporting them besides the authors' subjective experience.

I have seen the same idea with processes, pipelines, lists, bullet points, jsons, yamls, trees, prioritization queues all for LLM context and instruction alignment. It's like the authors take the structure they are familiar with, and go 100% in on it until it provides value for them and then they think it's the best thing since sliced bread.

I would like, for once, to see some kind of exploration/abalation against other methods. Or even better, a tool that uses your data to figure out your personal bias and structure preference for writing specs, so that you can have a way of providing yourself value.

hsaliak 50 minutes ago||
nobody knows what to build when everything can be built, there is no moat.
sunaookami 3 hours ago|||
It's Vibesmaxxing
IceDane 3 hours ago||
It's like horoscopes for the entirely-too-AI-pilled. Founded in nothing but vibes.

"Don't write prompts like that, do it like this! I swear it's better. Claude says so!"

wismwasm 4 hours ago||
Try openspec: https://github.com/Fission-AI/OpenSpec/blob/main/docs/gettin...
arikrahman 4 hours ago||
That was my initial thought when reading the headline but the author states they didn't know it existed before doing this project and critiques it.
brendanmc6 4 hours ago||
Indeed I have a lot of catch up to do, will spend some time with the popular tools before I go too much further down this road.
didgeoridoo 1 hour ago||
I’m building something similar with https://github.com/LabLeaks/special (apologies for the desultory slop-laden README, need to give that a lot more human attention) but I’ve gone in a slightly different direction: a “spec” is a product contract claim supported by attached tests that verify it. It’s a little Cucumber-y, if anyone remembers that, but a lot more flexible — you just write stuff like

  @spec LINT_COMMAND.ORPHAN_VERIFIES

  linter reports blocks that do not attach to a supported owned item.
Then

  #[test]
  // @verifies SPECIAL.LINT_COMMAND.ORPHAN_VERIFIES

  fn rejects_orphan_verifies_blocks() {
    let block = block_with_path("src/example.rs", &["@verifies EXPORT.ORPHAN"]);

    let parsed = parse_current(&block);

    assert!(parsed.verifies.is_empty());
    assert_eq!(parsed.diagnostics.len(), 1);
    assert!(
        parsed.diagnostics[0]
            .message
            .contains("@verifies must attach to the next supported item")
    );
}

And then the CLI command “special specs” pulls your specs and all attached verification + test code so you (or your LLM) to analyze whether the (hopefully passing!) test actually supports the product claim.

There’s also a bunch of other code quality commands and source annotations in there for architectural design & analysis, fuzzy-checking for DRY opportunities, and general codebase health. But on the overall principle, this article is dead-on: when developing with LLMs, your source of truth should be in your code, or at least co-located with it.

photios 4 hours ago||
Yesterday I heard about lat.md [1] which seems to have similar ideas about annotating code with spec refs. I now need to try them both.

[1] https://www.lat.md/

brendanmc6 4 hours ago||
Oh cool, thanks for sharing. LLMs tend to adopt these patterns on their own, so I'm sure my ideas are far from unique.
arikrahman 4 hours ago||
Agents.md actually scales and outperforms alternatives so I would be skeptical about overcomplicating the proces https://vercel.com/blog/agents-md-outperforms-skills-in-our-...
k9294 2 hours ago|
Small advice - make one repo “main” and link to it from the website instead of an organisation.

I wanted to star the project to track the progress but it feels a bit weird.. Which repo shall I track? Server? Cli? Sounds like a misc repos.

brendanmc6 2 hours ago|
Ooh thought I did that, thanks.
More comments...