Top
Best
New

Posted by kisamoto 10 hours ago

Reallocating $100/Month Claude Code Spend to Zed and OpenRouter(braw.dev)
226 points | 172 comments
wiether 9 hours ago|
People may feel differently about the fee that OpenRouter takes, but I think the service they provide is worth the extra cost.

Having access to dozens of models through a single API key, tracking cost of each request, being able to run the same request on different models and comparing their results next to each other, separating usages through different API keys, adding your own presets, setting your routing rules...

And once you start using an account with multiple users, it's even more useful to have all those features!

Not relying on a subscription and having the right to do exactly what you want with your API key (using it with any tool/harness...) is also a big plus to me.

therealpygon 6 hours ago||
I agree with you in certain circumstances, but not really for internal user inference. OpenRouter is great if you need to maintain uptime, but for basic usage (chat/coding/self-agents) you can do all of what you mentioned and more with a LiteLLM instance. The number of companies that send a bill is rarely a concern when it comes to “is work getting done”, but I agree with you that minimizing user friction is best.

For general use, I personally don’t see much justification as to why I would want to pay a per-token fee just to not create a few accounts with my trusted providers and add them to an instance for users. It is transparent to users beyond them having a single internal API key (or multiple if you want to track specific app usage) for all the models they have access to, with limits and logging. They wouldn’t even need to know what provider is hosting the model and the underlying provider could be swapped without users knowing.

It is certainly easier to pay a fee per token on a small scale and not have to run an instance, so less technical users could definitely find advantage in just sticking with OpenRouter.

BeetleB 3 hours ago|||
The two things I like about OpenRouter:

1. The LLM provider doesn't know it's you (unless you have personally identifiable information in your queries). If N people are accessing GPT-5.x using OpenRouter, OpenAI can't distinguish the people. It doesn't know if 1 person made all those requests, or N.

2. The ability to ensure your traffic is routed only to providers that claim not to log your inputs (not even for security purposes): https://openrouter.ai/docs/guides/routing/provider-selection...

It's been forever since I played with LiteLLM. Can I get these with it?

instalabsai 33 minutes ago|||
One additional major benefit of OpenRouter is that there is no rate limiting. This is the primary reason why we went with OpenRouter because of the tight rate limiting with the native providers.
napoleond 3 hours ago|||
> It doesn't know if 1 person made all those requests, or N.

FWIW this is highly unlikely to be true.

It's true that the upstream provider won't know it's _you_ per se, but most LLM providers strongly encourage proxies like OpenRouter to distinguish between downstream clients for security and performance reasons.

For example:

- https://developers.openai.com/api/docs/guides/safety-best-pr...

- https://developers.openai.com/api/docs/guides/prompt-caching...

BeetleB 3 hours ago||
Fair point. Would be good to hear from OpenRouter folks on how they handle the safety identifier.

For prompt caching, they already say they permit it, and do not consider it "logging" (i.e. if you have zero retention turned on, it will still go to providers who do prompt caching).

Deathmax 3 hours ago||
OpenRouter tells you if they submit with your user ID or anonymously if you hover over one of the icons on the provider, eg OpenAI has "OpenRouter submits API requests to this provider with an anonymous user ID.", Azure OpenAI on the other hand has "OpenRouter submits API requests to this provider anonymously.".
BeetleB 2 hours ago||
But does "anonymous user ID" mean that they make a user ID for you, and it's sticky? If I make a request today and another tomorrow, the same anonymous user ID is sent each time? Or do they keep changing it?
fg137 2 hours ago||||
> The number of companies that send a bill is rarely a concern

Not true in any non startup where there is an actual finance department

datadrivenangel 2 hours ago||||
LiteLLM had a major security incident recently, and often isn't actually that useful an abstraction...
cobertos 3 hours ago||||
Does OpenRouter perform better than LiteLLM on integration though? I found using Anthropic's models through a LiteLLM-laundered OpenAI-style API to perform noticably worse than using Anthropic's API directly. So I've scrapped considering LiteLLM as an option. It's also just a buggy mess from trying to use their MCP server. The errors it puts out are meaningless, and the UI behaves oddly even in the happy path (error message colored green with Success: prepended).

But if OpenRouter does better (even though it's the same sort of API layer) maybe it's worth it?

wongarsu 3 hours ago|||
A lot of inference providers for open models only accept prepaid payments, and managing multiple of those accounts is kind of cumbersome. I could limit myself to a smaller set of providers, but then I'm probably overpaying by more than the 5.5% fee

If you're only using flagship model providers then openrouter's value add is a lot more limited

rvnx 3 hours ago||
The main thing about Openrouter is also that they take 100% of the risk in case of overcharges from the models, you have an actual hard cap.

The minus is that context caching is only moderately working at best, rendering all savings nearly useless.

SR2Z 1 hour ago||
Is there any risk? Don't the model providers also bill by the token?
fuzzy2 1 hour ago||
The accounting could be asynchronous, so you could overshoot your budget by a few requests before you're blocked.
vidarh 2 hours ago|||
I love Openrouter. The ability to define presets, and the ease of access is well worth the few vs. juggling lots of providers separately. I maintain a few subscriptions too - including the most expensive Claude subscription - but Openrouter handles the rest for me.
r0fl 1 hour ago|||
Love openrouter I can use cheap models without having to have an api at a bunch of different providers and can use the expensive models when im in a pinch and am maxed out from claude or codex

well worth the 5% they take

spaniard89277 35 minutes ago|||
You can get the same with kilo gateway without the fee.
pixel_popping 8 hours ago||
Expect you don't have the right to do what you want with the API Key (see waves of ban lately, many SaaS services have closed because of it).
embedding-shape 8 hours ago||
Unless you provide some more details, at least outline what "do what you want" was in your case, this seems like just straight up FUD.
himata4113 7 hours ago||
openrouter accepts crypto so might have been some money laundering involved for reselling dirty crypto for llm api.

if that wasn't the reason, hey that's actually a great way to launder money (not financial advice).

embedding-shape 7 hours ago||
So you pay OpenRouter with cryptocurrencies, which they accept as a payment method, and then what, they block your account because the cryptocurrencies you paid with came from some account on the blockchain associated with other stuff?

Or what are you really saying here? I don't understand how that's related to "you don't have the right to do what you want with the API Key", which is the FUD part.

himata4113 7 hours ago||
You pay openrouter with dirty crypto, then you have a business which simply resells openrouter giving you clean fiat. I think openrouter specifically only banned those kind of accounts since that's what I have observed from other comments / research. numlocked in this thread has explicitely said that they don't ban accounts for any of the reasons specified above which narrows down the scope to some form of broken ToS specifically around fraud and money laundering.
embedding-shape 7 hours ago||
And then you go on HN and post "you don't have the right to do what you want"? Yeah, FUD and good riddance if so.
pixel_popping 6 hours ago|||
You are not allowed to resell Openrouter as an API yourself, so for example if you make a service that charge per token, you can't use Openrouter API for that, this is specified in their ToS, so no, you can't do what you want, what FUD?

Quote from their own TOS: access the Site or Service for purposes of reselling API access to AI Models or otherwise developing a competing service;

embedding-shape 6 hours ago||
Yeah, you're not allowed to do things that are specifically spelled out in the ToS, how is this surprising? Of course you don't get "unlimited access to do whatever you technically can", APIs never worked like that, why would they suddenly work like that?

When you say "you don't have the right to do what you want with the API Key" it makes it sound like specific use cases are disallowed, or something similar. "You don't have the right to go against the ToS, for some reason they block you then!" would have been very different, and of course it's like that.

Bit like complaining that Stripe is preventing you from accepting credit card payments for narcotics. Yes, just because you have an API key doesn't mean somehow you can do whatever you want.

pixel_popping 4 hours ago||
That's very different from the Stripe example, as opening a service like Openrouter isn't illegal, so that's only coming from it being opinionated, nothing to do with the law. And my example was for not so specific use cases but quite general one which is just to open let say a service like Opencode Zen and use Openrouter as a backend, this is explicity forbidden by Openrouter and it isn't against the law, that's not just a "niche use case".

Are we allowed yes or not to make a service that charge per Token to end-users, like giving access to Kimi K2.5 to end-users through Openrouter in a pay per token basis?

Vinnl 7 hours ago|||
That was a different user who wrote that.
embedding-shape 7 hours ago||
Yeah, I didn't mean them specifically, more a general "you".
Vinnl 23 minutes ago||
Ah fair enough.
supernes 8 hours ago||
On the topic of Zed itself as a VSCode replacement - my experience is mixed. I loved it at first, but with time the papercuts add up. The responsiveness difference isn't that big on my system, but Zed's memory usage (with the TS language server in particular) is scandalous. As far as DX goes it's probably at 85% of the level VSCode provides, but in this space QoL features matter a lot. Oh, and it still can't render emojis in buffers on Linux...
fishgoesblub 47 minutes ago||
I've been attempting to use Zed as a VSCode replacement but between the lack of bitmap font support (Criminal in an alleged code editor), and the weird UI choices, it's been hard. I want to love it, but what is "performance" if I have to spend more time working around the UI and lack of features from extensions. Strangest issue I've encountered is the colours being completely wrong when using Wayland.. colours are perfect when ran with Xwayland. I'll give it a plus though for native transparency for the background. Much nicer than having to make the entire window transparent, including the foreground text like with VSCode.
extr 3 hours ago|||
I actually find Zed pretty reasonable in terms of memory usage. But yeah, like you say, there are lots of small UX/DX papercuts that are just unfortunate. In some cases I'm not sure it's even Zed's fault, it's just years and years of expecting things to work a certain way because of VS Code and they work differently in Zed.

Eg: Ctrl+P "Open Fol.." in Zed does not surface "Opening a Folder". Zed doesn't call them folders. You have to know that's called "Workspace". And even then, if you type "Open Work..." it doesn't surface! You have to purposefully start with "work..."

udkl 2 hours ago|||
QoL features is where WebStorm shines! I don't look forward to when I have to open vscode instead sometimes.

Just the floating and ephemeral "Search in files" modal in Jetbrain IDEs would convince me to switch from any other IDE.

fxtentacle 2 hours ago|||
my favorite is Ctrl+Shift+A which lets you search through all available UI actions (hence the A). That's just so helpful if you know the IDE can do something but you forgot where in the menu structure it was. And to top things off, you can also use Ctrl+Shift+A to look up the keyboard shortcuts for every possible action
BoorishBears 1 hour ago|||
I still debate how much productivity I've gained from better AI compared to the loss from switching off WebStorm

But their tab complete situation is abysmal, and Supermaven got macrophaged by Cursor

thejazzman 3 hours ago|||
I think there’s a bug? It used to be memory efficient and now I periodically notice it explodes. Quit and restart fixes it

I don’t have any extensions installed and I’m basically leaving it open, idle, as a note scratch space. I do have projects open with many files but not many actual files are open

Anyway idk

tuzemec 7 hours ago|||
I have 4-5 typescript projects and one python opened in Zed at any given time (with a bunch of LSPs, ACPs, opened terminals, etc.) and I see around 1.2 - 1.4gb usage.

I opened just one of the typescript projects inside VSCode and I see something like 1gb (combining the helpers usage). I'm not using it actively, so no extra plugins and so on.

That's on mac, so I guess it may vary on other systems.

rzkyif 7 hours ago||
Same here: I found the multibuffers feature really useful, but the extension system really couldn't hold a candle to VS Code at the time of my testing

Spent a couple of hours trying to make the Svelte extension ignore a particular type of false positive CSS error, failed, and returned to VS Code

Will definitely give it another chance when the extension system is more mature though!

extr 3 hours ago||
I think you are kidding if you think you are going to be remotely approximately the quantity/quality of output you get from a $100/max sub with Zed/Openrouter. I easily get $1K+ of usage out of my $100 max sub. And that's with Opus 4.6 on high thinking.
alexjplant 17 minutes ago||
For personal use I've noticed Claude (via the web-based chat UI) making really bizarre mistakes lately like ignoring input or making completely random assumptions. At work Claude Code has turned into an absolute dog. It fails to follow instructions and builds stuff like a lazy junior developer without any architecture, tests, or verification. I don't know what they did but Anthropic's quality lead has basically evaporated for me. I hope they fix it because I've since adapted my project's Claude artifacts for use with Codex and started using it instead - it feels like Claude Code did earlier this year.

I'd like to give the new GLM models a try for personal stuff.

lelanthran 3 hours ago|||
> I easily get $1K+ of usage out of my $100 max sub. And that's with Opus 4.6 on high thinking.

And people keep claiming the token providers are running inference at a profit.

gruez 3 hours ago||
>And people keep claiming the token providers are running inference at a profit.

Not everyone gets $1K of usage, and you don't know how fat the per-token margins are. It's like saying the local buffet place is losing money because you eat $100 worth of takeout for $30.

lelanthran 24 minutes ago|||
> Not everyone gets $1K of usage, and you don't know how fat the per-token margins are.

Well, we're going to find out sooner rather than later. Right now you don't know how thin (or negative) the margins are, either, after all.

All we know for certain is how much VC cash they got. Revenue, spend, profit, etc calculated according to GAAP are still a secret.

kitsune1 2 hours ago|||
[dead]
Aurornis 2 hours ago|||
Some of the newer models available on OpenRouter are good, but I agree that none of them are a replacement for Opus 4.6 for coding.

If you're trying to minimize cost then having one of the inexpensive models do exploratory work and simple tasks while going back to Opus for the serious thinking and review is a good hybrid model. Having the $20/month Claude plan available is a good idea even if you're primarily using OpenRouter available models.

I think trying to use anything other than the best available SOTA model for important work is not a good tradeoff, though.

mikeocool 1 hour ago|||
Yeah — I just created an anthropic API key to experiment with pi, and managed to spend $1 in about 30 minutes doing some basic work with Sonnet.

Extrapolating that out, the subscription pricing is HEAVILY subsidized. For similar work in Claude Code, I use a Pro plan for $20/month, and rarely bang up against the limits.

causal 1 hour ago|||
And it scales up - the $200 plan gets you something like 20x what the Pro plan gets you. I've never come close to hitting that limit.

It's obviously capital-subsidized and so I have zero expectation of that lasting, but it's pretty anti-competitive to Cursor and others that rely on API keys.

walthamstow 56 minutes ago|||
I ran ccusage on my work Max account and I spend what would cost $300 a week if it was billed at API rates.
nothinkjustai 3 hours ago||
Not everyone is just vibecoding everything and relying on agents running sota models to do anything tho.
KronisLV 1 hour ago||
I tried using OpenRouter for the same kind of development I now do with Anthropic's subscription across Sonnet/Gemini/GPT models and it ended up being 2-3x more expensive than the subscription (which I suspect is heavily subsidized).

It's nice that it works for the author, though, and OpenRouter is pretty nice for trying out models or interacting with multiple ones through a unified platform!

ElFitz 9 hours ago||
Has anyone (other than OpenClaw) used pi? (https://shittycodingagent.ai/, https://pi.dev/)

Any insights / suggestions / best practices?

jsumrall 3 hours ago||
Its really fantastic. I can't imagine why you'd go through the effort using Claude Code with other models when pi is a much better harness. There's tons of extensions already available, and its trivial to prompt an LLM to create your any new extension you want. Lacking creativity and want something from another harness?

> Run <other harness> in tmux and interrogate it how feature X works, then build me the equivalent as a pi extension.

Maybe in a few years there will be obvious patterns with harnesses having built really optimal flows, but right now it works so much better to experiment and try new approaches and prompts and flows, and pi is the easiest one to tweak and make it your own.

ElFitz 1 hour ago||
> but right now it works so much better to experiment and try new approaches and prompts and flows, and pi is the easiest one to tweak and make it your own.

That’s what really appeals to me. I’ve been fighting Claude Code’s attempts to put everything in memory lately (which is fine for personal preferences), when I prefer the repo to contain all the actual knowledge and learnings. Made me realise how these micro-improvements could ultimately, some day, lead to lock-in.

> Run <other harness> in tmux and interrogate it how feature X works, then build me the equivalent as a pi extension.

I’ll give it a try!

simgt 8 hours ago|||
Yes, it's super cool. Check Mario's latest talk: https://www.youtube.com/watch?v=Dli5slNaJu0 Armin also has some videos covering it on his channel: https://www.youtube.com/@ArminRonacher/ Pi's Discord is still nice, even though it was a bit flooded after the openclaw thing.
WhyNotHugo 2 hours ago|||
Pi is a lot simpler than Claude and a lot more transparent in how it operates.

It's designed to be a small simple core with a rich API which you can use for extensions (providing skills, tools, or just modifying/extending the agent's behaviour).

It's likely that you'll eventually need to find extensions for some extended functionality, but for each feature you can pick the one that fits your need exactly (or just use Pi to hack a new extension).

dust42 7 hours ago|||
I really love it. The simplicity is key. The first play project I made with it was a public transport map with GTFS data - click on a stop and get the routes and the timetables for the stop and the surrounding ones. I used Qwen3.5-35B on Mac M1 Max with oMLX. It wrote 98% of the code with very little interaction from me. And very useful is the /tree feature to go back in history when the model is on a wrong track or my instructions where not good enough. I usually work in a two path approach: first let the model explore what it needs to fulfill the task and write it into CONTEXT.md (or any other name to your liking). Then restart the session with the CONTEXT.md. That way you are always nicely operating in 5-15k context, i.e. all is very fast. Create an account for pi (or docker) and make sure it can't walk into other directories - it has bash access. Add the browser-tools to the skills and load them when useful: https://github.com/badlogic/pi-skills

No need for database MCP, I use postgres and tell it to use psql.

Occasionally I use prettier to remove indentation - the LLM makes a lot less edit errors that way. Just add the indent back before you commit. Or tell pi to do it.

hayd 3 hours ago|||
had been using claude max/opus with pi and the results have been incredible. Having pi write an AGENTS.md and dip your feet into creating your own skills specific to the project.

With the anthropic billing change (not being able to use the max credits for pi) I think I have to cancel - as I'm whirring through credits now.

Going to move to the $250/mo OpenAI codex plan for now.

zweicoder 3 hours ago||
I was looking into this as well since Claude models are costing too much with the Extra Usage changes.

Is OpenAI codex not also charging by usage instead of subscription when using pi?

ElFitz 1 hour ago||
pi is what OpenClaw runs on, and so far OpenAI seems committed to it. No telling how long it will last.
sonar_un 2 hours ago|||
I use Daniel Meissler’s PAI and it’s been an incredible harness.
nocobot 8 hours ago|||
i really have been enjoying pi a lot

at first i thought i was goring to build lots of extra plugins and commands but what ended up working for me is:

- i have a simpel command that pulls context from a linear issue

- simple review command

- project specific skills for common tasks

Daviey 7 hours ago||
Reluctantly, the dev seems to have a stinky attitude.

He went on an "OSS vacation", which is perfectly reasonable and said he'd be back on a certain date. I had a PR open for a trivial fix, someone asked when it would land. I shared he was still away. After his return I politely asked, "@badlogic hey, what can we do to progress this? Thanks x"

I then got what I would consider an abusive reply, because he confused me with someone else. In the meantime he extended his vacation. Didn't even think his shitty attitude was worthy of an apology, that HE confused me with someone else.

https://github.com/badlogic/pi-mono/discussions/1475#discuss...

And another other thing I fixed with no attribution, just landed it himself separately. https://github.com/badlogic/pi-mono/discussions/1080

and

https://github.com/badlogic/pi-mono/issues/1079#event-223896...

Now he's seemingly marked anything with my name on as a "clanker", despite all my changes being by hand.

I've been around open source enough to have a thick skin, but when i'm doing something "for fun" and someone treats you like that, i'd rather avoid it as far as possible. I certainly could not in good faith use this project for anything work related.

crashprone 2 hours ago|||
You seem to have posted your polite question as a reply to the bot comment which talks about PR #1484 and not your PR. I'd say it's pretty obvious why the maintainer thought you were pushing the bot's PR.

As someone else pointed out cooler heads and less passive aggressive responses would've resolved this issue easily.

embedding-shape 7 hours ago||||
> Honestly, it seems you are grumpy, so it was probably a good idea to extend that vacation. Being rude just creates a more toxic environment for everyone. Maybe extend that break for the rest of the month and come back nicer? Thanks

Honestly, it seems like both of you were feeling a bit "grumpy" at the moment, but sending passive aggressiveness towards the maintainer you are trying to get to merge your code (or not your code, someone else's code?) seems like a very bold strategy regardless.

Daviey 6 hours ago||
You know, when I wrote that I genuinely meant it, or at least I think i did. It wasn't supposed to be passive aggressive. :(

But that doesn't negate the maintainer talking to people like that (and taking contributions without attribution).. and the net result is I don't want to use the software, and frankly they probably won't miss me.. so the end result is neutral.. I just find it sad.

raincole 2 hours ago|||
> Maybe extend that break for the rest of the month and come back nicer?

Quite sure most (perhaps >99%) adult people would consider this passive aggressive.

But yeah, I agree with you for the rest part. Why did Mario assume that bot is you...?

ZeWaka 1 hour ago||
if a human showed up directly under some bot bullshit pinging me I'd assume they were the bot operator as well
sodacanner 3 hours ago|||
I'm not sure how you'd be able to interpret that as anything other than passive aggressive.
NwtnsMthd 3 hours ago|||
But hey, the dev was generous to give it an MIT license, you could always just fork it and what you like ¯\_(ツ)_/¯
simlevesque 57 minutes ago||
I really don't like OpenCode. One thing that really irritated me is that on mouse hover it selects options when you're given a set of choices.
bashtoni 9 hours ago||
After hitting Claude limits today I spent the afternoon using OpenCode + GLM 5.1 via OpenRouter and I was very impressed.

OpenCode picked up my CLAUDE.md files and skills straight away, and I got similar performance to Opus 4.6.

sourcecodeplz 4 hours ago||
How much did it cost for how long?
bashtoni 32 minutes ago|||
~$1/hr over 4 hours.

I'm pretty conservative when it comes to clearing the context, and I also tend to provide the right files to work on (or at least the right starting point).

I had seen prior to using the model that it starts producing much worse results when the context used is larger, so my usage style probably helps getting better results. I work like this with Claude Code anyway, so it wasn't a big change.

BeetleB 3 hours ago|||
https://z.ai/subscribe

Many of us got the annual Lite plan when they had the $28 discount. But even at $120 I think it's a good deal.

jml78 1 hour ago||
I am trying to take this in the more giving way possible, anyone remotely considering that subscription should go on reddit and see all the people experiencing outages constantly and insanely slow speeds when it does work.

I have been wanting to subscribe but based on how awful the experience is for most people, I just can’t pull the trigger

BeetleB 1 hour ago|||
At $84, I can understand not taking the risk. But for $28 ... it was worth it.

FWIW, I've never dealt with outages since I signed up over 3 months ago (Lite plan). It is slow - always has been. I can live with that.

At the same time, I'm not using it for work. It's for the occasional project once in a while. So maybe I just haven't hit any limits? I did use it for OpenClaw for 2-3 weeks. Never had connection issues.

Looking at https://docs.z.ai/devpack/faq

you can see the details of their limits. Seems GLM 5.1 has low thresholds, and will get lower starting May. On Reddit I see some people switching to GLM 5 and claiming they haven't hit limits - the site doesn't indicate the limits for that model.

They also say that those who subscribed before February have different limits - unsure if it's lower or higher!

GLM-4.7 is still a fairly capable model. Not as good as Opus, but for most personal projects it's been adequate. I see on Reddit plenty of people plan using GLM-5.1, and use 4.7 for implementation.

srslyTrying2hlp 47 minutes ago||
[dead]
reddec 7 hours ago||
My 50c - ollama cloud 20$. GLM5 and kimi are really competitive models, Ollama usage limits insane high, no limits where to use (has normal APIs), privacy and no logging
my002 1 hour ago||
Interesting. I've always been turned off by how vague the descriptions of Ollama's limits are for their paid tiers. What sort of work have you been doing with it?
yieldcrv 2 hours ago||
yeah? why do you like that over using GLM5 in a VPS that charges by token use? $20 still cheaper and seamless to set up? how are the tokens per second?
hybirdss 36 minutes ago||
The bursty usage pattern is what kills the subscription value. I hit limits during mid-refactor and there's nothing to do but wait. The worst part is knowing the unused hours during the day are just gone.

OpenRouter credit rollover is the real insight — credits that don't expire vs time windows that reset whether you used them or not. I'm surprised Anthropic hasn't offered a token pack option alongside the subscription.

delduca 8 hours ago|
I also dropped Claude Code Max.

I switched to OpenCode Zen + GitHub Copilot. For some reason, Claude Code burns through my quota really quickly.

https://opencode.ai/zen

100ms 2 hours ago||
I dislike how Zen (and many similar cases, not picking on Zen here) report being not for profit or transparent, while the auto-recharge mechanism guarantees they are sitting on a float of at least $5 per account, and presumably an average of at least $10. That's something like 50 cents of interest income per year per account. It's not nothing and it's hardly egregious fraud, but I feel if they will do this when it's obvious what they're doing, what other corners might they cut

Honesty as a marketing strategy is really undervalued in cases like this

pprotas 1 hour ago|||
Yeah man, it's a grand scheme to skim 50 cents off you per year. All combined, that might be just enough to cover their website hosting costs.
vixalien 31 minutes ago|||
honestly the issue with Zen is that they collect and might sell your data
woutr_be 8 hours ago||
How does Claude Code compare to OpenCode Zen? I’m on the $20/month Claude plan, and was considering OpenCode Zen as well.

Due to the quota changes, I actually find myself using Claude less and less

delduca 7 hours ago|||
I mostly use Opus via Copilot with opencode, and I'll tell you, in the past few days, I've had long sessions (almost the whole day) without hitting rate limits. That's very different from Claude Code, which used to rate-limit me before even halfway through the day.
woutr_be 6 hours ago||
Just cancelled my Claude plan, so that I can switch over when it expires in a week. The usage limits somehow just make me less productive with it.
criley2 8 hours ago|||
I haven't tried $20 claude code recently, but I've used OpenCode Zen primarily so I can play with opensource/chinese models which are very inexpensive. I'd spend $0.50-$1.00 on a single claude opus 4.6 plan mode run, then have a chinese model execute the plan for like $0.10-$0.15 total. I'd keep context short, constantly start new threads, and get laser focused markdown plans and knowledgebase to be token efficient.

If I just let opencode zen run claude opus to plan and execute, I'd spend $20 in like 5 minutes lol

sourcecodeplz 4 hours ago||
Which chinese models do you use and do you use any for specific tasks?
msh 2 hours ago|||
kimi k2.5 works quite well and is super fast. Much faster than opus but not quite at the same quality level.
criley2 3 hours ago|||
Whenever a new one comes out, there's a good chance they're free for a week on Zen, so I try out any free ones. For example, MiniMax M2.5 and Qwen 3.6+ are free right now.

Personally, I've had a lot of good results in my little personal projects with Kimi K2.5, GLM 5 and 5.1, and MiniMax M2.5.

More comments...