I’ve been noticing that this simple reality explains almost all of both the good and the bad that I hear about LLM-based coding tools. Using AI for research or to spin up a quick demo or prototype is using it to help plot a course. A lot of the multi-stage agentic workflows also come down to creating guard rails before doing the main implementation so the AI can’t get too far off track. Most of the success stories I hear seem to be in these areas so far. Meanwhile, probably the most common criticism I see is that an AI that is simply given a prompt to implement some new feature or bug fix for an existing system often misunderstands or makes bad assumptions and ends up repeatedly running into dead ends. It moves fast but without knowing which direction to move in.
This is a real problem when the "direction" == "good feedback" from a customer standpoint.
Before we had a product person for every ~20 people generating code and now we're all product people, the machines are writing the code (not all of it, but enough of it that I will -1 a ~4000 line PR and ask someone to start over, instead of digging out of the hole in the same PR).
Feedback takes time on the system by real users to come back to the product team.
You need a PID like smoothing curve over your feature changes.
Like you said, Speed isn't velocity.
Specifically if you have a decent experiment framework to keep this disclosure progressive in the customer base, going the wrong direction isn't a huge penalty as it used to be.
I liked the PostHog newsletter about the "Hidden dangers of shipping fast", I can't find a good direct link to it.
https://newsletter.posthog.com/p/the-hidden-danger-of-shippi...
I suppose there is an argument that if you are building the wrong thing, build it fast so that you can find out more quickly that you built the wrong thing, allowing you to iterate more quickly.
A lot of people are so enamored by speed, they are not even taking the time to carefully consider the full picture of what they are building. Take the HN frontpage story on OpenCode: a maintainer admitted they keep adding many shallow features that are brittle.
Speed cannot replace product vision and discipline.
It also moves fast with a tendency to pick the wrong direction (according to the goal of the prompter) at every decision point (known or unknown).
Really like the article I think it is awesome, and I strongly believe AI for coding will stay, but I also beleive that we need to still have a strong understanding of why we are building things and what they look like.
An example being the common attitude that [advanced tech] is just a math problem to be solved, and not a process that needs to play itself out in the real world, interacting with it and learning, then integrating those lessons over time.
Another way to put this is: experience is undervalued, and knowledge is overvalued. Probably because experience isn’t fungible and therefore cannot be quantified as easily by market systems.
1. Probably not his original idea, and now that I think about it this is kind of more Hegelian. I’m not familiar enough with Hegel to reference him though.
I've been using AI to help me write it and I've come to a couple conclusions:
- AI can make working PoCs incredibly quickly
- It can even help me think of story lines, decision paths etc
- Given that, there is still a TON of decisions to be made e.g. what artwork to use, what makes sense from a story perspective
- Playtesting alone + iterating still occurs at human speed b/c if humans are the intended audience, getting their opinions takes human time, not computer time
I've started using this example more and more as it highlights that, yes, AI can save huge amounts of time. However, as we learned from the Theory of Constraints, there is always another bottleneck somewhere that will slow things down.
Coming up with a genuinely interesting gameplay loop with increasing difficulty levels and progressively revealed gameplay mechanics is a fascinating and extremely difficult challenge, no matter how much AI you throw at the problem.
„I was able to vibecode those 5 apps I always wanted but never had time to code them myself … it is so different now because — I don’t have time to use them”.
Lost me in paragraph three. We pay for those things because they're recognizable status symbols, not because they took a long time to make. It took my grandmother a long time to knit the sweater I'm wearing, but its market value is probably close to zero.
The fact that those items took a long time to make is part of what makes them status symbols though, because if you pay a lot of money for something that took no time to make at all (see most NFTs) you look like an idiot to a lot of people.
This sort of thing was done at a time when everybody did it, and now that it's not done, nobody does it
No kid ever said "did you see the sweater that Timmy's grandma knitted for him? That kid is so cool! "
Mostly because they all had grams sweaters as well.
I don't know what term you were looking for, but a handmade present for someone dear is about the furthest thing from a "status symbol" that I can think of:
- it can't be bought
- it can't be transferred without losing almost all value (ie: it's only valuable to you, or at most your family, eBay doesn't want it)
- it provides no improvement whatsoever in one's social standing
I can't connect it at all to your listed points. An Olympic medal is about obvious a status symbol as I can imagine but it can't (meaningfully) be bought or transferred.
The status signified with a knit sweater is membership (and good standing!) in a caring family with elders not yet fully subsumed into their phones.
People, acquaintances and strangers alike, frequently comment on the knit socks I often wear, ask after who made them, and all of a sudden we're on "how's your mom" terms.
https://www.ebay.com/b/Olympic-Medal/27291/bn_55191416?_sop=...
> People, acquaintances and strangers alike, frequently comment on the knit socks I often wear,
Ok, that explains pretty much everything about your line of thought.
Thanks.
> https://www.ebay.com/b/Olympic-Medal/27291/bn_55191416?_sop=...
Of course you can buy an Olympic medal. You can't buy the status conferred by the medal (of Olympic champion / nth runner up).
> Ok, that explains pretty much everything about your line of thought.
I don't understand this either. Are you insulting me?
I'm also completely unimpressed by someone wearing a Rolex though, so different mileage for different people.
Understanding words does not require being impressed by anything, nor caring about the opinion of kids.
If people don't consider that someone with more money is of a higher status then symbols of that wealth aren't meaningful.
I think a lot of people have an ingrained belief that "more money == more status"
You fill a jar with sand and there is no space for big rocks.
But if you fill the jar with big rocks, there is plenty of space for sand. Remove one of the rocks and the sand instantly fills that void.
Make sure you fit the rocks first.
You fill the bottle with half of the water, you put the fish in, you can fill in the other half. If you start with the first half, you will end up with more water.
Yes, there are bad metaphors, and people who take metaphors too seriously. That you can conjure a bad metaphor with somewhat similar to semantics to some other metaphor does not mean that said metaphor is bad.
That water overflow step is missing / implicit. But that's an observable event.
then you fill 3 liter bottle again, and pour the contents into the 5 liter bottle until the 5 liter one is full
empty the 5 liter bottle, and pour the 1 liter in the 3 liter bottle into the 5 liter bottle
fill the 3 liter bottle again and pour that into the 1 liter already in the 5 liter bottle to get 4 liters of water
> Given a 3-liter container and a 5-liter container, both initially empty, and access to tap water, how can you measure exactly 4 liters of water without using any additional containers
I've offered and received some convoluted metaphors recently, love leaning hard into this one.
Not sure, I used to be better at diagnosing this type of episode.
Agentic coding very much feels like a "video game" in the sense of you pull the lever and open the loot box and sometimes it's an epic +10 agility sword and sometimes its just grey vendor trash. Whether or not it generates "good" or even "usable" code fades to the background as the thrill of "I just asked for a UI to orchestrate micro services and BLAMMO there it was!" moves to the fore.