Posted by derek 7/1/2025
For example, I'm currently experimenting with an agent workflow for stock research. I've set up two AI roles: a 'Bullish Guy' and a 'Bearish Guy' and have them debate the pros and cons of a specific stock. The premise is that through this adversarial process, the AIs are forced to research opposing viewpoints, leading to a more comprehensive understanding and a superior final analysis. The idea was inspired by the kinds of arguments you see on social media.
it’s really promising, but I found focusing on a single task and doing it well is still more efficient for now. excited for where this goes
In case the author is lurking, you may want to apply the same fix they do in clojure-mcp: https://github.com/bhauman/clojure-mcp/blob/8150b855282babcd...
The insight that team had was that LLMs get confused with parens, but they are excellent at indentation, so if you run parinfer over the LLMs output it will be correct in 99% of cases.
What sort of subscription plan is that?
ccusage shows me getting over 10x the value of paying via API tokens this month so far...
npx ccusage@latest
Outputs a table of your token usage over the last few days, which it reads from the jsonl files that Claude Code leaves tucked away in the ~/.claude/ directory.I'll probably became a farmer by then.
It probably works well for small inputs and tasks well-represented in the training data (like writing code for well-represented domains).
But how does this work for old code, large codebases, and emergencies?
- Do you still "learn" the system like you used to before?
- How do you think of refactoring if you don't get a feel for the experience of working through the code base?
Overall: I like it. I think this adds speed for code that doesn't need to be reinvented. But new domains, new tools, new ways to model things, the parts that are fun to a developer, are still our monsters to slay.
Have you actually tried Claude Code? It works pretty well on my old code, medium size SaaS codebase. I’ve had it build entire features end to end in (backend, front end, data migrations, tests) in one or two prompts.
This made me chuckle.
Perfect example of why heavily LLM-driven devs and processes might want to pick a popular programming language which the LLM had a ton of training data for. Or a strong point for specialized LLMs (e.g. here it could be a smaller/cheaper/faster Clojure-specialized model).
I agree with this and it aligns with the general opinion about what is the true value the SWE's bring to the table.
Why do I hear the words "technical debt"? More to the point, the risk I see with this approach is that the author would end hp throwing away working and well tested code to implement some minor change. This has an high risk of introducing many easily avoidable bugs.