It's great that they're wrong, though. How often do you use crypto directly to pay for your bills and groceries? Must be nice not needing a credit card anymore. Those interest rates and credit scores can be a killer!
You do you, and I'll do me. Perhaps spend more time coding which you say you like to do and less time sneering at people who use different tools.
You should pick some of the substance of the article to take issue with instead of jumping to be a victim. They clearly read it if they wrote it.
I know there are people who oppose, say, syntax coloring, and I think that’s pretty weird, but I don’t go out of my way to dunk on them.
Did you even read the blog post? The entire blog is dunking on people who use the tools.
What embarrassment? This post, like plenty of Evelyn’s writing, reached the front page of HN and has a ton of commentary in agreement. Your comment, on the other hand, was downvoted to the bottom of the thread.
The author also draws and publishes what is, in their own words, “pretty weird porn”.
You should voice your disagreement with the contents of the post and explain what they are, if that is what you’re feeling. Discussion is what HN is for. But to believe the author is or should be suffering any kind of embarrassment about this post is detached from reality.
Still don't get it. LLM outputs are nondeterministic. LLMs invent APIs that don't exist. That's why you filter those outputs through agent constructions, which actually compile code. The nondeterminism of LLMs don't make your compiler nondeterministic.
All sorts of ways to knock LLM-generated code. Most I disagree with, all colorable. But this article is based on a model of LLM code generation from 6 months ago which is simply no longer true, and you can't gaslight your way back to Q1 2024.
There is no value in randomly choosing an API that exists. There is value in choosing an API that works.
When LLM makes up an API that doesn't even exist it indicates that it's not tied to the reality of the task of fetching a working API, so filtering out nonexistent APIs will not help the other results match better. But yes, they'll compile.
LLMs can write truly shitty code. I have a <50% hit rate on stuff I don't have to rewrite, with Sketch.dev, an agent I like a lot. But they don't fail the way you or this article claim they do. Enough.
Second, speak for yourself, you have no clue about everybody's experience to make such a universal claim.
Lastly, the article talks about author's experience, not yours, so you're the only one who can gaslight the author, not the other way around
For example, just yesterday I asked an AI a question about how to approach a specific problem. It gave an answer that "worked" (it compiled!) but in reality it didn't really make any sense and would add a very nasty bug. What it wrote (It used a FrameUpdate instead of a normal Update) just didn't make sense on a basic level of how the framework worked.
This is my problem: not that people are cynical about LLM-assisted coding, but that they themselves are hallucinating arguments about it, expecting their readers to nod along. Not happening here.
You made a similar claim: LLMs invent APIs that don't exist
1. A type-strict compiler.
2. https://github.com/isaacphi/mcp-language-server
LLMs will always make stuff up because they are lossy. In the same way that if I ask you to list the methods for some random object lib you'd not be able to do that; you use the documentation to pull that up or your code-complete companion. LLMs are just getting the tools for that.
I know the author already addressed this, literally calling out HN by name, but I just don't get it. You don't even need agents (though I'm sure they help), I still just use regular ChatGPT or Copilot or whatever and it's still occasionally useful. You type in what you want it to do, it gives you code, and usually the code works. Can we appreciate how insane this would have been, what, half a decade ago? Are our standards literally "the magic english-to-code machine doesn't work 100% of the time, so it's total crap, utterly useless"?
I absolutely agree with the general thrust of the article, the overall sense of disillusionment, the impact LLM abuse is going to have on education, etc. I don't even particularly like LLMs. But it really does feel like gaslighting to the extent that when these essays make this sort of argument (LLMs being entirely useless for coding) it just makes me take them less seriously.
Indeed. This is how to spot an ideologue with an axe to grind, not someone whose beliefs are shaped by dispassionate observation.
> But this article is based on a model of LLM code generation from 6 months ago which is simply no longer true, and you can't gaslight your way back to Q1 2024.
You’re ahead of the curve and wondering why others don’t know what you do. If you’re not an AI company, a faang, or an AI evangelist you likely haven’t heard of those solutions.
I’ve been trying to keep up with AI developments, and only learned about MCP and agentic workflows 1-2 months ago and I consider myself failing at keeping up with cutting edge AI development
Edit:
Also six months ago is Q1 2025, not 2024. Not sure if that was a typo or a need to remind you at how rapidly this technology is iterating
There are oh-so-many issues with LLMs - plagiarism/IP rights, worsening education, unmaintainable code - this should be obvious to anyone. But painting them as totally useless just doesn't make sense. Of course they work. I've had a task I want to do, I ask the LLM in plain English, it gives me code, the code works, I get the task done faster than I would have figuring out the code myself. This process has happened plenty of times.
Which part of this do you disagree with, under your argument? Am I and all the other millions of people who have experienced this all collectively hallucinating (pun intended) that we got working solutions to our problems? Are we just unskilled for not being able to write the code quickly enough ourselves, and should go sod off? I'm joking a bit, but it's a genuine question.
It's annoying to have to hand-code that stuff. But without Copilot I have to. Or I can write some arcane regex and run it on existing code to get 90% of the way there. But writing the regex also takes a while.
Copilot was literally just suggesting the whole deserialization function after I"d finished the serializer, 100% correct code.
Now that everything is containerised and managed by docker style environments, I am thinking about giving SBCL another try, the end users only need to access the same JSON REST APIs anyway.
Everything old is new again =)
Also, someone mathematically proved that's enough. And then someone else proved it empirically.
There was an experiment where they trained 16 pigeons to detect cancerous or benign tumours from photographies.
Individually, each pigeon had an average 85% accuracy. But all pigeons (except for one outlier) together had an accuracy of 99%.
If you add enough silly brains, you get one super smart brain.
Seems like 40 years of effort making deterministic computing work in a non-deterministic universe is being cast aside because we thought nondeterminism might work better. Turns out, we need determinism after all.
Following this out, we might end up with alternating layers of determinism and nondeterminism each trying to correct the output of the layer below.
I would argue AI is a harder problem than any humans have ever tried to solve, how does it benefit me to make every mundane problem into the hardest problem ever? As they say on the internet ...and now you have two problems, the second of which is always the hardest one ever.
LLM outputs are deterministic. There is no intrinsic source of randomness. Users can add randomness (temperature) to the output and modify it.
> But this article is based on a model of LLM code generation from 6 months ago
There hasn't been much change in models from 6 months ago. What happened is that we have better tooling to sift through the randomly generated outputs.
I don't disagree with your message. You are being downvoted because a lot of software developers are butt-hurt by it. It is going to force a change in the labor market for developers. In the same way the author is butt-hurt that they didn't buy Bitcoin in the very early days (as they were aware of it) and missed the boat on that.
I made the same claim in a widely-circulated piece a month or so back, and have come to believe it was wildly false, the dumbest thing I said in that piece.
So far the only model that showed significant advancement and differentiation was GPT-4.5. I advise to look at the problem and read GPT-4.5 answer. It'll show the difference to other "normal models" (including GPT-3.5) as it shows considerable levels of understanding.
Other normal models are now more chatty and have a bit more data. But they do not show increased intelligence.