Posted by tanelpoder 8 hours ago
IE, using agents to iterate through many possible approaches, spike out migrations, etc might save a project a year of misadventures, re-designs, etc, but that productivity gain _subtracts_ the intermediate versions that _didn't_ end up being shipped.
As others have mentioned, I think yak-shaving is now way more automated. IE, If I want to take a new terminal for a spin, throw together a devtool to help me think about a specific problem better, etc, I can do it with very low friction. So "personal" productivity is way higher.
In that they obviously have no real utility, sure. There hasn't been a paradigm shift, they still suck at programming, and anyone trying to tell you otherwise almost certainly has something to sell you.
Maybe I should have said "obvious to me," but I guess I just struggle to see how a serious crack at using modern opus in claude code doesn't make it obvious at this point.
I'd really recommend trying the "spike out a self-contained minimal version of this rearchitecture/migration and troubleshoot it iteratively until it works, then make a report on findings" use-case for anyone that hasn't had luck with them thus far and is serious about trying to reach conclusions based on direct experience.
In the last few months, Gemini (and I) have written for highly personal, very niche apps that are perfect for my needs, but I would never dream of releasing. Things like cataloguing and searching my departed mom‘s recipe cards, or a text message based budget tracker for my wife and I to share.
These things would never be released or available as of source or commercial applications in the way that I wanted them, and it took me less time to have them built with AI then it would have taken me to Research existing alternatives and adapt my workflow/use case to fit whatever I found.
So yeah, there are more apps but I would venture to say you’ll never see most of them…
Yes, you do produce more code. But LoC produced is never a healthy metric. Reviewing the LLM generated code, polishing the result and getting it to production-level quality still very much requires a human-in-the-loop with dedicated time and effort.
On the other hand, people who vibe code and claims to be 10x productive, who produces numerous PRs with large diffs usually bog down the overall productivity of teams by requiring tenuous code reviews.
Some of us are forced to fast-track this review process so as to not slow down these "star developers" which leads to the slow erosion in overall code quality which in my opinion would more than offset the productivity gains from using the AI tools in the first place.
I think this might be more of an comment on software as a business than AI not coding good apps.
200 years ago text was much more expensive, and more people memorized sayings and poems and quotations. Now text is cheap, and we rarely quote.
And even “product engineers” often do not have experience going from zero to post sales support on a saas on their own.
It is a skill set of its own to make product decisions and not only release but stick with it after the thing is not immediately successful.
The ability to get some other idea going quickly with AI actually works against the habits needed to tough through the valley(s).
The tools for generating AI code aren't yet capable of producing code that is decent enough for general purpose use cases, with good robust tests, and clean and quality.