Posted by jakelsaunders94 8 hours ago
Asking it to draft was weakening my own skills.
As shown in "Normal Accidents" the strength is as high as its weaknesses, and in any complex system this is even more a problem. A catastrophic event is still to happen with AI as it happened in basically every complex system. They ocurred with trained people that wasnt believing in magic or laziness... so the scenario is even worse for AI.
Yes, I'm bored about people that believe in magic and the ghosts the are emerging and are yet to be seen.
All they essentially did was tell the LLM to test and verify whether the answer is correct with a prompt like the following:
>"You just edited X. Before moving on, verify the change is correct: write a short inline python -c or a /tmp test script that exercises the changed code path, run it with bash, and confirm the output is as expected."
Now whether this is true, I don't know, but I think talking about this kind of stuff is cool!
Endlessly grooming the Agent reminds me of Gastown.
Curios to see what he'll present, if, from his 700+ contributions in private repositories.
But nobody wants to hear about prompt calibration or pipeline architecture. They want to hear "I replaced my whole team with agents." The boring, useful work is invisible, and the flashy stuff gets all the oxygen
It will calm down once the dust starts to settle and there's some kind of consensus on how the chips have fallen.
Also there is an irony that talking about being sick of talking about AI is still talking about AI.
The only thing that triggers me about it peoples inability to understand how a scam works, after falling for such scams the n-th time.
Hyperloop, ubeam, blockchain, Elon musk taking all to mars....
In these line of scams, LLMs are a wet dream...