Top
Best
New

Posted by neilfrndes 6 hours ago

Claude for Small Business(www.anthropic.com)
272 points | 198 comments
CSMastermind 5 hours ago|
I'm increasingly convinced that there's a killer app waiting for whoever can come up with a UI that makes claude code or codex accessible to the average user.

Onboarding my non-software engineer teammates to it has super-charged them and essentially given them all their own personal developer that can automate tasks for them. Managing codebases, etc. is still a hassle though.

90% of the power of Excel was that it was functionally a database that a normal person could actually use. I think we'll see something similar with coding agents.

ageitgey 3 hours ago||
> that makes claude code or codex accessible to the average user

That's what they aim Claude Cowork at. Every executive/leader I've shown Claude Cowork to has gone from 'what is AI' to 'vibecoding whole apps' in weeks. Then when Claude is down for an hour, they get visibly angry and don't remember how to do anything pre-Claude :)

I understand the impulse to provide a UI to manage codebases, etc. But my observation is that these people just ask Claude to do whatever it is they need done. Codebase needs managing? They just ask Claude to do it. No idea how to deploy an app? They just ask Claude to do it.

Any app built on top of this stack to 'make it easier' is competing with 'I don't care what's happening, just ask Claude to do it'.

Ucalegon 1 hour ago|||
>Every executive/leader I've shown Claude Cowork to has gone from 'what is AI' to 'vibecoding whole apps' in weeks.

Do you, and those executives, own the risks associated with that practice? Are those risks actually indemnified?

Its neat that 'anyone can do anything' but if they don't actually know what the risk to business or 3rd parties, why is this a good thing, especially in the enterprise where there are actors who are explicitly looking for this type of environment to exploit?

ageitgey 1 hour ago|||
These are largely friends and peers, so they ultimately own their own risks. But I'm not saying it is good or bad. I'm just telling you what is happening in the real world. Every senior person I know, whether a high tech exec or a solo coffee bean importer, is vibing to some degree. Some will be more successful than others.

I've been working in tech since the late 90s. This is the biggest and most sudden change in company behavior I've ever seen. The only thing that comes close was the web 1.0 world in the 90s where everything suddenly became websites.

That creates tons of risks and opportunities. Good and bad. Maybe a great time to start a security company. But maybe a terrible time to be a small time web app developer when your clients can get 'good enough' in minutes for dollars on their own.

Ucalegon 1 hour ago||
>But I'm not saying it is good or bad.

Wait, you exposed people to a technology, taught them how to use it, then you are not going to own the implications of that action without teaching them about the risks or telling them how they need to ensure they don't shoot themselves in the face or violate their duty of care?

Do you understand what you are saying and the implications of that in the real world relative to the insurance contracts that they have?

Your company is associated with HIPAA, you should have a much higher standard than this.

tclancy 25 minutes ago|||
Play the ball, not the man, dude. Hectoring people on the Internet because you're stressed out about something isn't going to magically fix how you feel. Digging into their profile to make it personal is three steps too far.
Ucalegon 17 minutes ago||
We are talking about one person's introduction of a technology to persons and the implications of that action within the framework of enterprise governance and risk, it is one in the same. If anything, who a person is, their knowledge of the domain and the associated implications that action has on the domain has relevancy where someone who is ignorant of implications may have more grace than someone who has the experience to know better. The passive lack of accountability or responsibility relative to that does matter given the context.
ageitgey 37 minutes ago|||
You are assuming like 12 things that aren't true in this response.
Ucalegon 34 minutes ago||
Explicitly name them then.
baxtr 51 minutes ago|||
What kind of risk do you see?
Ucalegon 45 minutes ago||
Depends on what types of apps are being built, what data they touch, and what those apps are exposed to from a network perspective. Ie; all of the fundamentals of information/network security. Generally speaking, most executives do not have an information/network security background but do have privileged access to extremely valuable information, even if an attacker just has access to their email.
nicce 3 hours ago||||
> Then when Claude is down for an hour, they get visibly angry and don't remember how to do anything pre-Claude :)

The drug is scary when everyone is depending on it. I wonder what is future like.

puchatek 2 hours ago|||
The future is perpetually dealing with the fallout from all the vibe coding as the pool of people who'd have a shot at fixing it gets smaller and smaller. Shitty will be the new normal.
freetanga 1 hour ago|||
I feel like it will be like going back to the 80s, when PCs became a norm and most programmers and hobbyists could code without the need of a University or a Corporation. Thousands of shareware apps you had to navigate, everyone trying to solve the same problems from different angles..

I do agree quality will be missed, and shadow IT will be again a big issue like at the end of the 80s and early 90s.

ElFitz 1 hour ago|||
> Shitty will be the new normal.

I’ve heard the same from the best devs, and some who thought themselves to be the best, I’ve known long before LLMs were ever a thing.

I’m sure others heard the same when JavaScript and Python became near ubiquitous. When PHP emerged. When C supplanted Fortran and COBOL. When these two took over from Assembly. When punch cards went the way of the dodo.

There’s always someone for whom shitty is becoming the new normal. If that makes it a rule, what do we make of that rule?

microtonal 1 hour ago|||
There are different magnitudes of shitty.

Also we went from compilers with an IDE that had a debugger, profiler, built-in help and would fit on a 3.5" disk and would load on machines with 640KiB RAM (Turbo Pascal) to chat apps or password managers that are hundreds of megabytes and regularly gobble up more than a gigabyte of memory because they ship with their own browser.

Something is lost along the way.

sersi 1 hour ago||||
To be fair with how powerful our computers are, it's a pity that electron apps like bitwarden, spotify are so slow and consume so much resources. I do miss the time when a lot of apps were snappy
freetanga 1 hour ago||||
Maybe it’s a process. Many of the transitions you mentioned did bring shitty apps (not all of them, the ones replacing tech for tech were mostly ok, the ones democratizing dev did come with a quality drop), but eventually Darwinism will take effect and trim the long tail.

Coding per se is not hard. Proper engineering is. I do hope this change brings a change in focus (people train in algorithms, efficiency, solid development patterns) but I am afraid it won’t be the case.

Saline9515 1 hour ago|||
"With a punchcard at least, I can verify what the input is! Unlike those new 'transistors' that are so unreliable!"
mercanlIl 9 minutes ago||||
I think there are some pretty good ways to understand it now.

When the electricity goes out, (most) people get similarly upset. No electricity means no internet, and all of a sudden everything that people had planed to do can’t be done until the power returns.

tyre 3 hours ago||||
Same as anything else. It’ll go down sometimes, people will take a break and chat, then it will come back up.

Like Slack or GitHub or AWS or whatever. It’s almost always a net positive to wait vs do it yourself.

lukan 2 hours ago|||
I think the scenario was more of, if really everyone depends on claude, then better nothing critical(medical software, aviation, traffic controll ..) breaks while claude is offline.
andrewl-hn 42 minutes ago|||
At least some of the projects in these industries now specify strict no-AI-use policies in contracts. I participate in a few of these, and it’s becoming a bit of a pain, because all dev tool vendors insist on adding AI features, and if there’s no way to turn them off completely we have to migrate away.

However, the temptation of productivity gains are strong, and few of the customers look into relaxing these rules.

jebus989 1 hour ago|||
The good thing is we've learned this already from cloud. When one AWS region is degraded we all failover to other regions, and then other cloud providers, right? ...right?
vrganj 2 hours ago|||
I'm more scared at everyone outsourcing their thinking to a private, for-profit company.

What could possibly go wrong.

dns_snek 2 hours ago||
Thinking, yes, but also secrets, access and effective control of important services in every country and company worldwide, centralized in the US (or anywhere else) where the NSA can take the driver's seat at any time. "AI" is the ultimate sleeper agent.
ValentineC 2 hours ago||||
> The drug is scary when everyone is depending on it. I wonder what is future like.

I can't wait for a Hollywood blockbuster that'll pretty much be science non-fiction.

dheera 1 hour ago||||
> wonder what is future like

Probably "don't do anything to upset AI companies or you will effectively become a handicapped person"

Not that different from life in China: "don't do anything to upset Tencent and AliPay or you will become an outcast"

Or life in the US if you're a content creator: "don't do anything to upset Meta or Youtube or you will not be able to pay your rent"

The future: ToS basically becomes law, and you will be stripped of your own second brain if you violate it or say anything they deem "sensitive"

oulipo2 1 hour ago||||
Full of security holes
safety1st 3 hours ago||||
Seems far less scary to me than, say, building an electrical grid in a cold climate, where if it fails for a few days people start to die. Oh wait...
M95D 2 hours ago|||
Why would they die in cold climate? I would expect them to die in hot climate (no AC - heat stroke, no refrigerator - food poisoning), not the cold where they would have wood/gas heating.
coldtea 2 hours ago|||
It's the same, on steroids.
M95D 2 hours ago|||
Imagine what happens if computers stop working* and you have to go back to pen and paper for a few days.

* ransomware attack, fire in the server room, database HDD crash, car accident takes out the internet connection, ...

ElFitz 2 hours ago||||
> I understand the impulse to provide a UI to manage codebases, etc. […] 'I don't care what's happening, just ask Claude to do it'.

Reading the first part, I was going to say they don’t even care about whether or not there’s a codebase. It doesn’t matter; it could be all gremlins and hamsters in wheels for all they care, and for all they should care. All that matters is the functionality, the value it gives them.

We’re even getting disposable code now. Entire single-use ephemeral web apps, built on the go to enable, visualise, or simplify a specific thing, then thrown away.

Will it all lead to some trouble? Definitely. So did computers, and so did the internet.

Weird times. Fun times.

rahoulb 1 hour ago||
When I quit my day job and started Rails freelancing a big chunk of my work was from companies with "that tech guy" who had built a database in Microsoft Access that was vital to the department's operations. And then either left the company - or the app had started to fall apart under its own weight.

I would get called in to rewrite it, using a proper database, documented rules and ensure it stayed scalable - and everyone would be happy.

These Access "apps" were abominations from a technical point of view - but they got the job done without having to spend a load of money on off-the-shelf or bespoke software. And the "tech guy" made a valuable contribution to the company. It's only at a certain point that Access started to struggle.

I foresee the exact same thing happening in the near future - except we won't be building the replacement apps ourselves - we'll just know how to give the coding agents well-specified prompts and tell them when they're making a mistake.

mattmanser 1 hour ago||
But at least you could basically follow their logic.

I think what a lot of us are concerned about is that the vibe-coded stuff bloats fast. It's so verbose and all over the place, that picking that thing apart will be a huge job, and relying on an AI to pick apart work that an AI already failed to maintain seem like wishful thinking.

It's literally "The AI is failing! Don't worry I'll just use AI to fix the AI!".

rahoulb 1 hour ago|||
The worst I would ever get was "here's our Access database - can you rewrite it". That was utterly useless to me.

What I needed to do was sit with a user (not a manager/the person buying my services) and ask them to show me the different things they did with the software. Then I could write a spec for the actual _feature_ and would only need to look at the existing codebase if they needed data transferring across[1]. I don't see why our new LLM-based future would be any different

[1] Of course this meant I would leave out edge-cases and/or weird quirks of the system - often this was actually a bonus as they were either no longer relevant or worked that way because that was the only way they knew how to do it

sersi 1 hour ago|||
Yes, as long as context size increase and llm improve at least there's a way out through using AI but once the progress stops...
bandrami 1 hour ago||||
Yeah I'm realizing now how many of you guys work in industries with no data security/protection requirements
senexox 9 minutes ago|||
Exactly. The tools aren't the rate limiting factor for me. I can automate an entire department right now with Claude but I can't because of regulations and audits. Basically, turning an error prone manual process into a probabilistic process that Claude would do far more accurately in the end than what we do now. The process wouldn't be "repeatable" though by the letter of the regulation so would open the company up to automated regulatory violations and existential fines. The technical issues for me are trivial but the regulations are insurmountable. The bubble is in the TAM. My work is exactly who Claude for Small Business would be aiming at but we can't do anything with these tools because of regulation. That is a huge % of the economy.
newsclues 1 hour ago|||
There are requirements they just don’t get enforced enough to matter
morpheuskafka 2 hours ago|||
> Any app built on top of this stack to 'make it easier' is competing with 'I don't care what's happening, just ask Claude to do it'.

To put it another way, the customers of these frontier models are implicitly being competed against by the model itself.

rib3ye 55 seconds ago|||
Non-engineer average user here: this is what cursor is for!
insin 1 hour ago|||
I don't think it needs to specifically be a coding agent for the average user, creating apps for whatever they want to do, just something that can use code and has appropriate access for what they're already asking it to do (instead of the model bullshitting to them that it can do it, annoying them), and some way to make it repeatable when needed, like skills.

I'm currently doing something like this in the internal model-independent LLM chat app I work on at a F100, specifically targeted at our everyday users. <input type="file" webkitdirectory> lets the user give the model read and write access to a local folder (and OPFS lets us reuse the same fs tools we give the model for files manually attached to the chat, or for files tools want to create if they haven't granted folder access).

Every time we used to release a new version it was "still can't handle the 6MB Excel file I drop into it" when that was being extracted to CSV and added to context - now it can poke about in the big Excel file directly with SheetJS to pull the sheets/headers and inspect the shape of the data, and use locally sandboxed code execution to write code against either extracted data or the spreadsheet itself via SheetJS for pivot tables and such (all locally - none of which need go into the context).

The base models are good enough at tool calling (I really mean Claude, though, the GPTs just go on a tear calling tools with no context for the user) they're already decent at automating stuff for the user without a dedicated harness (our default system prompt is still "You are a helpful AI assistant", lol). Add tools for Graph API stuff, and now it can pull the nightly batch file from a support inbox, unzip the spreadsheet within, diff it against yesterday's and generate an import file for new users and draft an email to welcome them, something that used to be a daily support task (which I'd already automated most of - but now you don't need a dev for this kind of thing). Or go find the big 450,000+ row spreadsheet that's being automated somewhere on SharePoint, pull it down in 150,000 row chunks (Graph Excel REST API limit) and write code to go figure out whatever the user is asking.

Having implemented and used it, I like this setup so much it kinda ruined Claude.ai and ChatGPT.com for me, so I've hooked up similar access for them using a browser extension to add the folder picker input, with the extension talking to a local server to tell it which folder to give access to, and Claude/ChatGPT talking to the same server over MCP via a CloudFlare Tunnel to work with the selected folder.

disillusioned 2 hours ago|||
We're building something along these lines, but since our roots are a consulting business, we're still building around the idea that there needs to be an expert integrator doing the front-loading work of discovery/decomposition/scoring of tasks/implementing them as those agents. These tools are terrifying to anyone not quite technical, and it turns out, people are bad at decomposing their own work, let alone describing it in a box with a blinking cursor.

We're obviously going to be holding ourselves back in terms of scale and in terms of not being a "true" SaaS with this approach, but my thesis is that we get much higher quality results and higher compliance/activation and can charge more for the bespoke model backed by our own platform.

mettamage 2 hours ago|||
> I'm increasingly convinced that there's a killer app waiting for whoever can come up with a UI that makes claude code or codex accessible to the average user.

I haven't tried it, or know a lot about it, but isn't this the whole claw thing?

tgv 1 hour ago|||
True story, heard yesterday from a consultant who was working with some VP type (not a large company, but still high management): VP uploads a spreadsheet to Claude and tells it to remove column F.

The power of Excel is not what it was. Nor is the power of ordinary thought.

whiplash451 51 minutes ago|||
> a UI that makes claude code accessible

Isn’t that literally Claude’s web UI?

bstsb 46 minutes ago||
while there are some tools available for the web UI, like building small React apps or making diagrams, it doesn't have the same loop as Claude Code in terms of iteratively building or fixing
devsda 3 hours ago|||
> Onboarding my non-software engineer teammates to it has super-charged them and essentially given them all their own personal developer that can automate tasks for them.

This is probably fine as long as the code is acting on local resources. The moment you have vibe coded software interacting with shared state or database the risk increases exponentially and all it takes to have a bad day is a poorly worded prompt from one of those users.

Some oversight by humans or automated guardrails will probably reduce those instances.

eecc 3 hours ago||
> Claude, fix the bug. Make no mistakes.

/s

otabdeveloper4 2 hours ago||
You forgot to add "you are an expert software engineer with PhD level architecture insights".
devsda 1 hour ago||
haha..After all "prompt engineering" is the mystic art of magecraft that uses forbidden incantations to summon the souls of special experts and make them possess our computers to do our bidding.
sersi 1 hour ago||
Sometimes watered down though. When I summon the soul of Linus, he is nowhere near as scathing or biting as the original :)
robbomacrae 3 hours ago|||
I'm trying to do this with orcabot.com

A figma like dashboard for turning ClaudeCode, Gemini Cli, Codex into an OpenClaw but with security measures to break the lethal trifecta while running on a VM.

But it's not quite there in terms of usability. I agree that is the hardest part of the equation. It's something I'm constantly experimenting with and haven't found the solution to it yet. Open to feedback!

fooker 1 hour ago|||
Microsoft is trying this with copilot, but they are calling everything copilot so YMMV.
lanyard-textile 4 hours ago|||
I am building a product in that space :)

It's targeted for creatives atm. For the few in private testing, it's been amazing what they're able to do with the little tooling I've given them. It is a legitimate change in their daily drive.

operatingthetan 3 hours ago||
>I am building a product in that space :)

I don't know anyone not building a product in that space

lanyard-textile 3 hours ago|||
I think everyone is making bespoke versions of what they think people want. It all feels gimmicky and dev oriented.

I have a vision for what will be the next household ChatGPT:

1. An actually frictionless way of keeping the human in the loop. My product is primarily targeting that: Your tools should feel like an extension of you, not replacing you.

2. Juggling work. I feel like what I'm making here is the secret sauce, so keeping a hush on it :)

3. Keeping all your work in one place. Drawing, sketching, developing, emailing, planning, writing; there is no reason to depend on other apps if you have one place that does it all, and it's the best offering among them.

Edit with some follow up thoughts -- I think what I'm trying to make is best summarized as claude code for non-developers (that's what I put in my YC application), but I think what I'm trying to make doesn't quite even have a developer equivalent.

There's not an environment you can go into right now and say "after this builds every single time, deploy to this machine" and it actually seamlessly does that. The tech is there but making it a whole Factorio-esque operation is still very manual -- and that's what I'm solving.

lukan 2 hours ago||
"I feel like what I'm making here is the secret sauce"

Good for your feelings, but I feel the same for my work ..

The main problem is still, agents are not reliable and what normal (and dev) people really want, is to have them reliable. Or well, tools to manage unreliable agents in a more clear way.

lanyard-textile 2 hours ago||
;) Then I think I have the trillion dollar idea. We'll see. Good luck to you.
lukan 2 hours ago||
Same to you.

(It is a big market I think)

endofreach 3 hours ago|||
So, what are you building in that space?
Hamuko 3 hours ago|||
I wouldn't want to build a business that was so dependent on a massive third-party that can either cut off my access or copy my design at any time of their choosing.
PAndreew 2 hours ago||
I was thinking about this and there are several aspects that can still make this viable. 1) AI labs are incentivised to increase token consumption because literally that's their product. The only thing they sell AFIAK are tokens (and maybe a teensy bit of user data). So if you build a product that is actively reducing token consumption (which they simply cannot do without hurting themselves even if their marketing fluff says otherwise) you'll save large amounts of money for your customers and they'll choose you. 2) Big providers want to funnel every prompt into their servers. If you're in a regulated market or simply don't want to share every detail with an American or Chinese megacorp you are in trouble. BUT open weight models are now quite capable for "small business stuff" and they can be self hosted. If you can bundle this into your service, in other words actually care about their privacy, they will choose you. Even more so if you're in Europe.
brainless 1 hour ago|||
I really believe that the Spreadsheets UX is great for mainstream users and that is what drives me for my coding agent that uses the sheets UX: https://github.com/brainless/nocodo

Super early stage but I am really happy to read your comment.

ignoramous 4 hours ago|||
> whoever can come up with a UI that makes claude code or codex accessible to the average user

You mean UX? Isn't Claude Cowork supposed to be 'Claude but for normies'? As for Claude Code / OpenAI Codex for non-programmers, believe Replit, Loveable, & others are trying & succeeding.

WhatsApp comes to mind in how its sole focus on replacing SMS (rather than Skype/AOL/MSN Messenger/YChat/GChat) meant it had no (user-facing) password/username, no elaborate signup, no login, no chat/friend requests, no sync etc. & became the biggest social network right under the nose of well resourced competitors with worldwide distribution, like Google & Facebook.

pmontra 4 hours ago||
Business wise, neither Google nor Facebook were impacted IMHO. Google sells the tools that WhatsApp need to run and Facebook bought WhatsApp and kept its FB users in house.

Probably phone operators were not impacted too: SMSes bundled with flat plans are still flat plans and Europe style unlimited calls + 100 SMS per month plans are still there and those SMSes are still mostly unused.

So we could have a killer app and yet nothing changes in the flow of money around it.

UX wise, WhatsApp is a big improvement over SMS. Vocal messages, I'm not a fan of them. A waste of my time.

graemep 2 hours ago||
Google was impacted: their chat product is pretty much dead.

Mobile network operators lost the profits (at prices that were pretty much pure margin) they had on pay as you go messages, and messages not included in flat plans (e.g. overseas SMS's). They also lost a huge amount on highly profitable overseas calls. Those of us with family in other countries save a lot of money by using Whatsapp and similar instead of phone calls.

dnnddidiej 2 hours ago|||
Lovable?
vasco 3 hours ago|||
Whoever does it everyone else will just prompt the same UX.
yordan_kavalov 2 hours ago|||
Yes, totally agree. Spent a few years in operations consulting and our clients' people were doing such amounts of mind-numbing repetitive work you wouldn't believe. Funny thing is, they are so used to it, they don't realize how wasteful it is. Yet, they are "afraid" of AI and new technologies in general, because it is something new and unfamiliar. However, when you show them something simple, e.g. how to write an Excel formula, they feel extremely motivated and empowered. So yes, if anyone can make AI feel less "scary" and approachable so that ordinary non-tech-savvy people can click around and see how they can automate some basic stuff, it will make them feel they have superpowers.
edf13 4 hours ago|||
[dead]
LPisGood 4 hours ago|||
I was just thinking about that earlier this week.

Claude can write code pretty well, but there are just a few tasks that I need to do to orchestrate everything. If it could do those tasks well even some of the time it would be about 10x more useful.

olliem36 4 hours ago|||
I agree and that's what i'm working on (for businesses) - an all-one-one consolidated AI application that's setup and ready for non-technical users.

It's called Zenning AI - we're a small team in London, testing it with a few companies at the moment!

dbuxton 4 hours ago||
We’re (harriethq.com) trying to do this by reframing it as a “provisioning” challenge - how do you get your connectors installed on non-technical desktops, how do you give some easy pre-bake recipes that wake them from their dogmatic slumber

Honestly though we are finding that a little FDE to set up pre-bake stuff that’s sufficiently specific to the customer is needed. Otherwise people are like, “I don’t need to close the books, I need to do a per-working-day profitability analysis for 10 EU countries with different public holidays”, and they get stuck there.

hommelix 5 hours ago||
By coincidence, I've looked yesterday a small documentary [1] about the people tagging all those invoices to train theses models. For 120 €/month they are reading about 1000 to 4000 invoices per day and check and tag them for AI training.

[1] https://www.arte.tv/en/videos/126831-000-A/arte-reportage/

cantalopes 4 hours ago||
Reminds me of openai paying Kenyans $2/hr to flag violent and toxic stuff for them and a bunch of people ending up with ptsd
iamflimflam1 4 hours ago|||
https://www.theguardian.com/technology/2023/aug/02/ai-chatbo...
hommelix 3 hours ago||||
In that video over Madagascar, the lowest tier jobs on AI tagging is at 1 €/3h of tagging, beating the Kenyan price.
madbkarim 4 hours ago|||
Source? Curious to know more.
iamflimflam1 4 hours ago|||
https://www.thebrink.me/the-ghosts-in-the-machine-inside-ai-...
intended 4 hours ago||||
> https://www.businessinsider.com/openai-kenyan-contract-worke...

> https://www.wsj.com/tech/chatgpt-openai-content-abusive-sexu...

> https://www.youtube.com/watch?v=qZS50KXjAX0

> https://www.bbc.com/news/av/world-africa-66514287

> https://www.vice.com/en/article/openai-used-kenyan-workers-m...

esseph 4 hours ago|||
There's tons of articles all over Google about this, it's not exactly hidden knowledge hoarded by this HN poster.

Example: https://www.theguardian.com/world/2024/dec/18/why-former-fac...

sph 1 hour ago|||
AI: Actual Indians^WMalagasy
wiseowise 4 hours ago|||
> For 120 €/month they are reading about 1000 to 4000 invoices per day and check and tag them for AI training.

AGI will solve poverty, btw. Any second now. Just need 500 bil more bro.

Barbing 5 hours ago|||
Were they sore about it?

Or don’t tell me, if it’s well worth the 24min watch

hommelix 3 hours ago||
Oh no! The ones working at 120€/month are the happy few. This is above mid range income in Madagascar. I just wanted to point out that this is not all automated running on GPUs. There are people involved, more than I thought before viewing this video.
elric 2 hours ago||
OCR based invoice recognition has been a solved problem for well over a decade. Source: I've consulted for a company doing that. No exploitation. No LLMs. Just clever engineering.

In my neck of the woods, B2B invoices are now required to be delivered over the Peppol network in UBL format, which further improves reliability.

Doesn't necessarily eliminate the need for an accountant, because the chosen UBL standard has lots of room for interpretation and ambiguity, and it's impossible to uniformly decide how process an invoice based on the invoice alone (e.g. is this deductible? is this even a business expense at all? which ledger should this go in? etc).

invoicenav 1 hour ago||
[flagged]
fnoef 4 hours ago||
You are absolutely right. I shouldn’t have paid that invoice from ScamInc. Would you like me to help you file for bankruptcy?
arjie 6 hours ago||
I understand why this is a good idea. I have Claude Code hooked up to my mail synced via IMAP, my Mercury read-only token, and beancount, and it gets almost all of my invoices and categorizes them. The tedious portion for a lot of this is:

* find invoice I_E for expense E

* associate and categorize E based on I_E and transaction field

These things are annoying but Claude Code is great at it and it leaves a much smaller set I have to manually resolve. This is a class of problems that are tractable and checkable, which I happily use LLMs on. If it miscategorizes it, I'm going to see it because I'm looking over the accounts. In fact, I was previously using a different accounting app which had poor API support, so I dumped it so I could use Claude and it's incredible how much this helps me.

There is an enormous number of use-cases that Claude/GPT are good for and the hard part is market penetration here. As an example, my dad was looking at some statistical health survey data in India and working out what things you could glean from it. Claude identified the things that would complicate his analysis in no time. He's 70 years old, and he'd done it all manually until he asked me (I've got a Mathematics degree) if something made statistical sense to do. I told him what it likely was and then asked him to try Claude. Knocked out his work and mine in moments. But he didn't think to use it. Now I have to get him a ChatGPT/Claude subscription.

It's like how if you go to the Datadog pricing page they don't list a feature set. They have all these use-case lists with prices. You can build things using their base metrics functionality and logs functionality but showing the use-cases must have more adoption.

Barbing 5 hours ago|
>[on] the Datadog pricing page…showing the use-cases must have more adoption.

Interesting, sometimes they want to show you they’ll simply charge 2-3 percent of your monthly spend (https://www.datadoghq.com/pricing/?product=audit-trail#produ...)

camel_gopher 4 hours ago||
2-3 percent; so far (Homer Simpson)
amelius 8 minutes ago||
Was hoping for an on-premises solution. Sending your data and your clients' data to a cloud is unacceptable for many small businesses.
trumbitta2 1 hour ago||
Let me get this straight: a few times per month, someone posts horror stories about how Claude led to losing data and money.

Anthropic's response: let's make a nice package out of this, and let's target specifically the businesses that are less likely to be ready to manage such horrible events.

Ucalegon 1 hour ago||
The reality is, for a lot of people, they do not care about risk or implication or cost, as so long as they see things moving forward, especially if they do not understand what they are dealing with. The desire of 'build, build, build', to these people does not have a downside because they do not have the knowledge of what the implications of that actually means nor is there a culture associated with the duty of care that should come with the liability associated with other people's data.

Also, small business contracts likely do not have the same type of language around indemnity/SLAs, so it is easier for the harms of this type of system to go unpunished because those who are harmed are even less knowledgeable.

sofixa 1 hour ago|||
Don't forget Microsoft researchers finding that multi-agent, multi-tool workflows result in at least 20% of the original content getting corrupted in the chain: https://www.theregister.com/ai-ml/2026/05/11/microsoft-resea...
dude250711 1 hour ago||
"someone..." with enough social media weight that is.

It's just like getting Google support.

tim-projects 3 hours ago||
> Claude helps take the late-night work off their plates.

This is dangerous. Relying on so much of your business on a third party. We've seen this many times before where businesses get destroyed because something gets broken somewhere that they have outsourced and have no control over.

In my view this service should not be used, unless there is a local llm or clear manual alternative.

Then the question begs - Why use Claude at all?

Maybe a proof of concept only while you come up with a real solution. Maybe to use claude to get rid of Claude

The people who get dazzled by bright lights are going to be the ones licking their wounds later. There is going to be eggs on faces one day.

mhitza 2 hours ago||
> D.3. Limitations of Outputs; Notice to Users. It is Customer’s responsibility to evaluate whether Outputs are appropriate for Customer’s use case, including where human review is appropriate, before using or sharing Outputs. Customer acknowledges, and must notify its Users, that factual assertions in Outputs should not be relied upon without independently checking their accuracy, as they may be false, incomplete, misleading or not reflective of recent events or information. Customer further acknowledges that Outputs may contain content inconsistent with Anthropic’s views.

Must be nice being able to ruthlessly lie with "this is the future" marketing claims, while hiding behind this term of service.

behaviors 2 hours ago||
It is a far bit tougher to actually get the clankers to speak accurately. I understand the legal perspective, with OpenAI talking about depression use cases, these companies who are running computers for users have to worry the software might harm the user(through themselves) and the leagl fallout needs protected.

It amazes me that we are going to litigate this like they did with cars over horses, or machines vs human labor. I honestly don't think Claude should be running companies.

grumbelbart 3 hours ago||
Of course, should it be as cost efficient as claimed and if you don't use it but everybody else does, you might be pushed out of the market.
jryio 6 hours ago||
I run a s business (small if you compare it to tech companies).

I can tell you the drag is between your own tools and the real world (which is very messy and inconsistent): taxes, compliance, payroll, amendments, share structures, etc.

Within my island, my books are in order, invoices and time keeping is fully automated, calendars and sales pipelines are connected.

I'm sure there are many businesses whose inner islands are not as orderly. The zillion tools out there all try to bring equanimity to the chaos and yet here we still are with fresh books, quickbooks, and xero...

cik 5 hours ago||
A deacde ago Xero, Shoeboxed, Calendly, Payment Evolution, and a time tracker eliminated all my overhead.

I scaled to 30+ people with automated administration. My cost was under $150 a month for everything we needed to run a successful consultancy and product business. Our accountant was blown away by how simple his life was.

I'm constantly amazed at how it has gotten much worse in the resulting decade.

jorisw 5 hours ago||
How did it get worse?
hirako2000 4 hours ago||
Wrappers around LLMs promise to bridge that gap. I'm sure it can do well for the vast majority of cases. But I do wonder what the outliers would cost.

E.g traditional automation + humans handling the drag = $4,000 per month with a couple of known blunder each year

vs traditional automation + AI = $400, with unknown number of blunders.

Of course it depends how much a blunder costs, to solve, or swallow. But I would bet that accounting errors even for a small business would cost the business on the long run. And that's assuming we don't yet have adversarial behavior which we can expect to come from both the inside and the outside.

SoftTalker 6 hours ago||
Waiting to hear the stories of things Claude did running amok in Quickbooks.
jrickert 3 hours ago||
I’ve given it access to my small business books for the last few months (attended sessions only) and so far it’s helped me clean up countless errors made by humans, at the expense of a small handful of duplicated transactions that got shaken out pretty quickly.
elric 2 hours ago||
How do you know those duplicates are the only errors it made? You weren't aware of the apparently countless human errors before, so how would you be aware of Claude's errors?
bontaq 6 hours ago||
It's a fascinating angle they've taken to give Claude your payroll. I guess we've reached this part of the AI race and they're running ahead of people realizing what it can do.
borski 4 hours ago||
Preparing payroll is different from running payroll. A human should still have to review it, as it’s the person running it (and the employer) that’s liable.
sourcecodeplz 5 minutes ago|
"From these tools, it can plan payroll, close the month, run a sales campaign, chase invoices, and more." wow
More comments...