Top
Best
New

Posted by svara 19 hours ago

Ask HN: How is AI-assisted coding going for you professionally?

Comment sections on AI threads tend to split into "we're all cooked" and "AI is useless." I'd like to cut through the noise and learn what's actually working and what isn't, from concrete experience.

If you've recently used AI tools for professional coding work, tell us about it.

What tools did you use? What worked well and why? What challenges did you hit, and how (if at all) did you solve them?

Please share enough context (stack, project type, team size, experience level) for others to learn from your experience.

The goal is to build a grounded picture of where AI-assisted development actually stands in March 2026, without the hot air.

305 points | 494 commentspage 8
brunooliv 9 hours ago|
Couldn’t read the entire comments but, my experience has been overwhelmingly positive so far. I think what helps me be effective is a combination of factors: I work only in a modern, well-documented and well-architected Java codebase with over 80% test coverage.

I only use Claude Code with Opus 4.6 on High Effort.

I always, ALWAYS treat my “new job” as writing a detailed ticket for whatever it is I need to do.

I give the model access to a DB replica of my prod DB that I create manually.

I do NOT waste time with custom agents, Claude.md files or any of that stuff.

When I put ALL of the above together, the results ARE THE PROMISED LAND: I simply haven’t written a single line of code manually in the last 3 months.

wasabinator 8 hours ago|
I find this pretty interesting. I am curious though: Did you dislike coding? You sound genuinely excited to not be doing it anymore.

For me I have been a coder since a very young age and I am nearing the end of my career now. I still love writing code to problem solve just as much as the first day I learnt to code. The thought of something taking that task away from me doesn't fill me with glee.

A parallel for me is if I enjoyed puzzle pages and those brought me with joy and satisfaction employing my grey matter to solve, I just wouldn't find it interesting to have an agent complete the forms to me, with me simply guiding the agent to clues.

cadamsdotcom 12 hours ago||
Models aren’t reliable, and it’s a bottleneck.

My solution was to write code to force the model down a deterministic path.

It’s open source here: https://codeleash.dev

It’s working! ~200k LOC python/typescript codebase built from scratch as I’ve grown out the framework. I probably wrote 500-1000 lines of that, so ~99.5% written by Claude Code. I commit 10k-30k loc per week, code-reviewed and industrial strength quality (mainly thanks to rigid TDD)

I review every line of code but the TDD enforcement and self-reflection have now put both the process and continual improvement to said process more or less on autopilot.

It’s a software factory - I don’t build software any more, I walk around the machine with a clipboard optimizing and fixing constraints. My job is to input the specs and prompts and give the factory its best chance of producing a high quality result, then QA that for release.

I keep my operational burden minimal by using managed platforms - more info in the framework.

One caveat; I am a solo dev; my cofounder isn’t writing code. So I can’t speak to how it is to be in a team of engineers with this stuff.

cluckindan 11 hours ago|
My most productive day last week was a net of -10k lines (yes, minus ten thousand).

No AI used.

cadamsdotcom 10 hours ago||
Congratulations, honestly, but I would not do that for a job.

Metaphorically speaking, you’re out there sprinting on the road while people who’ve made agentic coding work for them are sipping coffee in a limo.

People who haven’t made agentic coding work (but do it anyway) are sipping coffee in the back of a limo that has no brakes. No thanks to that.

ramoz 12 hours ago||
Right now I enjoy the labs' cli harnesses, Claude Code, and Codex (especially for review). I do a bunch of niche stuff with Pi and OpenCode. My productivity is up. Some nuances with working with others using the same AI tools- we all end up trying to boil the ocean at first- creating a ton of verbose docs and massive PRs, but I/they end up regressing on throwing up every sort of LLM output we get. Instead, we continously refine the outputs in a consumable+trusted way.

My workday is fairly simple. I spend all day planning and reviewing.

1. For most features, unless it's small things, I will enter plan mode.

2. We will iterate on planning. I built a tool for this, and it seems that this is a fairly desired workflow, given the popularity through organic growth. https://github.com/backnotprop/plannotator

  - This is a very simple tool that captures the plan through a hook (ExitPlanMode) and creates a UI for me to actually read the plan and annotate, with qol things like viewing plan diffs so I can see what the agent changed.
3. After plan's approved, we hit eventual review of implementation. I'll use AI reviewers, but I will also manually review using the same tool so that I can create annotations and iterate through a feedback loop with the agents.

4. Do a lot of this / multitasking with worktrees now.

Worktrees weren't something I truly understood the value of for a while, until a couple weeks ago, embarrassingly enough: https://backnotprop.com/blog/simplifying-git-worktrees/

jsunderland323 9 hours ago||
I've been working on a thing for worktrees to work with docker-compose setups so you can run multiple localhost environments at once https://coasts.dev/. It's free and open source. In my experience it's made worktrees 10x better but would love to hear what other folks are doing about things like port conflicts and db isolation.
queenkjuul 12 hours ago||
Worktrees slap.
jellyfishbeaver 12 hours ago||
Same attitudes as others in this thread.

For personal projects and side company, I get to join in on some of the fun and really multiply the amount of work I can get through. I tend to like to iterate on a project or code base for awhile, thinking about it and then tearing things down and rebuilding it until I arrive at what I think is a good implementation. Claude Code has been a really great companion for this. I'd wager that we're going to see a new cohort of successful small or solo-founder companies that come around because of tools like this.

For work, I would say 60% of my company's AI usage is probably useless. Lots of churning out code and documents that generate no real value or are never used a second time. I get the sense that the often claimed "10x more productive" is not actually that, and we are creating a whole flood of problems and technical debt that we won't be able to prompt ourselves out of. The benefit I have mostly seen myself so far is freeing up time and automating tedious tasks and grunt work.

0x6d61646f 11 hours ago||
I had automation setup for anything I needed for work, gen AI made me feel like I had to babysit a dumb junior developer so I lost interest

Managment uses it to make mock websites then doesn't listen when we point out flows, so nothing new there

Some in digital marketing are using it for data collection/anlysis, but it reaches wrong conclusions 50% of the time (their words) so they are slowly dropping it and using it for meneal tasks and simple automations

In design we had a trial period but has the same issue as coding: either it makes something a senior designer could have made in 2 minutes or it introduces errors that take a long time to fix, to then do it again the next prompt

we are a senior dev team, although relative small, and to me it seems like it only really works as a subsitute for junior devs... but the point of junior devs is to grow someone into a senior with the knowledge you need in the company so i don't really get the usecase overall

crsl 13 hours ago||
At work I mostly use claude code and chatgpt web for general queries, but cursor is probably the most popular in our company. I don’t think we are "cooked" but it definitely changes how development will be done. I think the process of coming up with solutions will still be there but implementation is much faster now.

My observations:

1. What works for me is the usual, work iteratively on a plan then implement and review. The more constraints I put into the plan the better.

2. The biggest problem for me is LLM assuming something wrong and then having to steer it back or redoing the plan.

3. Exploring and onboarding to new codebases is much faster.

4. I don’t see the 10x speedup but I do see that now I can discard and prototype ideas quickly. For example I don’t spend 20-30 minutes writing something just to revert it if I don’t like how it looks or works.

5. Mental exhaustion when working on multiple different projects/agent sessions is real, so I tend to only have one. Having to constantly switch mental model of a problem is much more draining than the “old” way of working on a single problem. Basically the more I give in into vibing the harder it is to review and understand.

Tade0 16 hours ago||
Daily Claude user via Cursor.

What works:

-Just pasting the error and askig what's going on here.

-"How do I X in Y considering Z?"

-Single-use scripts.

-Tab (most of the time), although that doesn't seem to be Claude.

What doesn't:

-Asking it to actually code. It's not going to do the whole thing and even if, it will take shortcuts, occasionally removing legitimate parts of the application.

-Tests. Obvious cases it can handle, but once you reach a certain threshold of coverage, it starts producing nonsense.

Overall, it's amazing at pattern matching, but doesn't actually understand what it's doing. I had a coworker like this - same vibe.

eranation 15 hours ago|
Opus 4.5 max (1m tokens) and above were the tipping point for me, before that, I agree with 100% of what you said.

But even with Opus 4.6 max / GPT 5.4 high it takes time, you need to provide the right context, add skills / subagents, include tribal knowledge, have a clear workflow, just like you onboard a new developer. But once you get there, you can definitely get it to do larger and larger tasks, and you definitely get (at least the illusion) that it "understands" that it's doing.

It's not perfect, but definitely can code entire features, that pass rigorous code review (by more than one human + security scanners + several AI code reviewers that review every single line and ensure the author also understands what they wrote)

shireboy 11 hours ago||
I am a developer turned (reluctantly) into management. I still keep my hands in code and work w team on a handful of projects. We use GitHub copilot on a daily basis and it has become a great tool that has improved our speed and quality. I have 20+ years experience and see it as just another tool in the toolbox. Maybe I’m naive but I don’t feel threatened by it.

At least at my company the problem is the business hasn’t caught up. We can code faster but our stakeholders can’t decide what they want us to build faster. Or test faster or grasp new modalities llms make possible.

That’s where I want to go next: not just speeding up and increasing code quality but improving business analytics and reducing the amount of meetings I have to be in to get business problems understood and solved.

GMoromisato 9 hours ago||
I'm working on a startup, mostly writing C++, and I'm using AI more and more. In the last month I have one machine running Codex working on task while I work on a different machine.

I have to think like micro-manager, coming up with discrete (and well-defined) tasks for the AI to do, and I periodically review the code to make it cleaner/more efficient.

But I'm confident that it is saving me time. And my love for programming has not diminished. I'm still driving the architecture and writing code, but now I have a helper who makes progress in parallel.

Honestly, I don't want to go back.

davemp 10 hours ago|
Professionally, sending our code off prem is not an option. Frankly I don’t understand why executives are okay with AI companies training LLMs on their IP. Unless they own a significant stake in the AI company I guess.

Personally, it’s been decent for generating tedious boilerplate. Though I’m not sure if reading the docs and just writing things myself would have been faster when it comes time to debug. I’m pretty fast at code editing with vim at this point. I’m also hesitant to feedback any fixes to the AI companies.

I’ve found “better google” to be a much more comfortable if not faster way to use the tools. Give me the information, I’ll build an understanding and see the big picture much better.

More comments...