Top
Best
New

Posted by lumpa 12/22/2025

A year of vibes(lucumr.pocoo.org)
196 points | 125 commentspage 2
summarity 12/22/2025|
Here’s something else that just started to rally work this year with Opus 4.5: interacting with Ghidra. Nearly every binary is now suddenly transparent, in many cases it can navigate binaries better than source code itself.

There’s even a research team that has bee using this approach to generate compilable C++ from binaries and run static analysis on it, to find more vulnerabilities than source analysis without involving dynamic tracing.

theptip 12/22/2025||
A really interesting point that keeps coming up in discussions about LLMs is “what trade-offs need to be re-evaluated”

> I also believe that observability is up for grabs again. We now have both the need and opportunity to take advantage of it on a whole new level. Most people were not in a position where they could build their own eBPF programs, but LLMs can

One of my big predictions for ‘26 is the industry following through with this line of reasoning. It’s now possible to quickly code up OSS projects of much higher utility and depth.

LLMs are already great at Unix tools; a small api and codebase that does something interesting.

I think we’ll see an explosion of small tools (and Skills wrapping their use) for more sophisticated roles like DevOps, and meta-Skills for how to build your own skill bundles for your internal systems and architecture.

And perhaps more ambitiously, I think services like Datadog will need to change their APIs or risk being disrupted; in the short term nobody is going to be able to move fast enough inside a walled garden to keep up with the velocity the Claude + Unix tools will provide.

UI tooling is nice, but it’s not optimized for agents.

shimman 12/22/2025|
Do you have any example repos of these OSS projects? I'm being reminded of this post every time people keep extolling how "productive" LLMs are:

https://mikelovesrobots.substack.com/p/wheres-the-shovelware...

Where is the resulting software?

theptip 12/23/2025|||
https://tools.simonwillison.net/
adamisom 12/22/2025||||
>Where is the resulting software?

Everywhere.

Remember Satya Nadella estimating 30% of code at Microsoft was written by AI? That was March. At this point it's ubiquitous—and invisible.

oblio 12/23/2025|||
> At this point it's ubiquitous—and invisible.

So, basically UFOs since smartphones put a camera in everyone's pocket.

bonzini 12/22/2025|||
> Everywhere.

Show the PRs.

pigpop 12/22/2025|||
Here you go https://ampcode.com/news/social-coding
shimman 12/22/2025||
I don't think using an AI company that relies on selling AI solutions to make money as a unbiased source of info, but what do I know? I'm not a VC investor and skeptical of the rich + elites in general.

Usually the best rule of thumb is to be against anything these people are for.

pigpop 12/22/2025||
Do you want me to link you directly to the repos and threads on that page, I can if you can't find them.
shimman 12/23/2025||
No, I'm good. I do find it telling they don't list one of the more public LLM adoptions being forced fed to MSFT workers:

https://github.com/dotnet/runtime/pull/115762

pigpop 12/23/2025||
Ok, I can see why they wouldn't list a copilot related stuff since that page is just providing a list of public Amp threads, some of which have public repos. Copilot is a different framework so you wouldn't find it on that page. I linked to the Amp threads because you can see both the conversation as well as the repo for those that include one.
rootnod3 12/22/2025||
"I have seen some people be quite successful with this."

Wait until those people hit a snafu and have to debug something in prod after they mindlessly handed their brains and critical thinking to a water-wasting behemoth and atrophied their minds.

EDIT: typo, and yes I see the irony :D

comex 12/22/2025||
Just be glad that there remains a concrete benefit to not atrophying your mind and deeply understanding your code. For now. In the long run, I suspect the behemoth will become just as capable at debugging and dealing with complexity as humans. At that point, human involvement in the actual code will be pointless, and the only remaining human skill needed will be properly directing the agents – the skill those people are learning right now.

(I don’t relish this future at all, myself, but I’m starting to think it really will happen soon.)

rootnod3 12/23/2025||
The only future I see is that prompts will become so refined that you give them all the requirements an they write the code for ya. And that prompt language has to be unambiguous and so detailed that you loop around and re-invented a programming language, just this time it wastes tons of water.
wiseowise 12/22/2025||
> Wait until those people hit a snafu and have to debug something in prod after they mindlessly handed their brains and critical thinking to a water-wasting behemoth and atrophied their minds.

You've just described typical run of the mill company that has software. LLMs will make it easier to shoot yourself in the foot, but let's not rewrite history as if stackoverflow coders are not a thing.

rootnod3 12/22/2025|||
Difference: companies are not pushing their employees to use stack overflow. Stack overflow doesn't waste massive amounts of water and energy. Stack overflow does not easily abuse millions of copyrights in a second by scraping without permission.
rootnod3 12/22/2025|||
Another difference: stack overflow tells you you are wrong or tells you and do your own research or to read the manual (which in a high percentage of cases is the right answer). It doesn't tell you that you are right and proceeds to hallucinate some non-existent flags for some command invocation.
pigpop 12/22/2025|||
This is a problem but it's a known one which both Google and Anthropic seem to be making progress towards solving. I've had a full on argument with Gemini 3 where it turned out I was wrong and it correctly stuck to its guns and wouldn't let me convince it otherwise. It eventually got through to me about the mistake I made and I learned something useful from it. Sonnet and Opus are still a bit too happy to tell you "you're absolutely right" but I've noticed more pushback creeping in in the right places. It's a tough balance to get right, nobody wants to pay for a service that just tells them "no" whenever they want to try something silly or unconventional.
christophilus 12/22/2025|||
It mostly incorrectly flags your question as a dup.
tjr 12/22/2025|||
There have been lots of tools and resources that have promised (and delivered!) increased programming productivity.

Individual results may vary, but it seems credible that thoroughly learning and using an editor like Vim or Emacs could yield a 2x productivity boost. For the most part, this has never really been pushed. If a programmer wanted to use Nano (or Notepad!), some may have found that odd, but nobody really cared. Use whatever editor you like. Even if it means leaving a 2x productivity boost on the table!

Why is it being pushed so hard that AI coding tools in particular must be used?

rootnod3 12/22/2025|||
I am not contesting that stackoverflow isn't bad in many regards, but to equate that to massive PRs or code changes done via AI slop is a different level. At worst, you might get a page or two out of stack overflow but still need to stitch it together yourself.

With LLMs you can literally ask it to generate entire libraries without activating a single neuron in your nogging. Those two do NOT compare in the slightest.

rootnod3 12/22/2025||
Sorry, but why would including the prompt in the pull request make any difference? Explain what you DID in the pull request. If you can't summarize it yourself, it means you didn't review it yourself, so why should I have to do it for you?
theshrike79 12/22/2025|
You're making assumptions, of course you add BOTH.

The point of adding the "prompt", or the discussion with the LLM is learning. You can go back and see what was the exact conversation.

rootnod3 12/23/2025||
Sounds more like just adding a ton of wasted time for the reviewer to read through those discussions. At least summarize it yourself, e.g. "After discovering manpage XYZ, it became clear that the correct usage of this function is fooBar()".
theshrike79 12/23/2025||
Why would the reviewer need to read through discussions? The description + code should be just fine.

It's like having someone watch a livestream screen recording of you writing the code.

It's nice to have there IF you need to go back and learn something, but hardly a review requirement.

bigfishrunning 12/23/2025||
> My biggest unexpected finding: we’re hitting limits of traditional tools for sharing code. The pull request model on GitHub doesn’t carry enough information to review AI generated code properly — I wish I could see the prompts that led to changes. It’s not just GitHub, it’s also git that is lacking.

I find when submitting a complex PR, i tend to do a self review, adding another layer of comments above those that are included in the code. Seems like a nice place to stuff prompts

netdevphoenix 12/23/2025||
In case people don't know, the author of the post is the creator of Flask, possibly Python's most popular micro web framework.
petcat 12/22/2025||
I respect Armin's opinions on the state-of-the-art in programming a lot. I'm wondering if he finds that "vibe coding" (or vibe engineering) is particularly pleasant and effective in Rust compared to, say, Python.
johnwheeler 12/22/2025|
I bet it would be probably even nicer. I've been programming DSP in C++ with JUCE. I have a very rusty C++ experience from years ago, but it's getting me through a lot of it, and I feel pretty comfortable. Maybe my ignorance is bliss, and I'm really just putting out bad shit.
adamisom 12/22/2025||
The very first thing I did vibe-coding was commit my prompts and AI responses. In Cursor that's extremely easy—just 'export' a chat. I stopped for security concerns but perhaps something like that is the way.
anshulbhide 12/22/2025||
> The pull request model on GitHub doesn’t carry enough information to review AI generated code properly — I wish I could see the prompts that led to changes. It’s not just GitHub, it’s also git that is lacking.

Yes! Who is building this?

amarant 12/22/2025|
Create a folder called "prompts". Create a new file for each prompt you make, name the time after timestamp. Or just append to prompts.txt

Either way, git will make it trivial to see which prompt belongs with which commit: it'll be in the same diff! You can write a pre-commit hook to always include the prompts in every commit, but I have a feeling most Vibe coders always commit with -a anyway

NitpickLawyer 12/22/2025||
It's not just that. There's a lot of (maybe useful) info that's lost without the entire session. And even if you include a jsonl of the entire session, just seeing that is not enough. It would be nice to be able to "click" at some point and add notes / edit / re-run from there w/ changes, etc.

Basically we're at a point where the agents kinda caught up to our tooling, and we need better / different UX or paradigms of sharing sessions (including context, choices, etc)

JKCalhoun 12/22/2025|
Got distracted: love the "WebGL metaballs" header and footer on the site.