Posted by david927 1 day ago
Ask HN: What Are You Working On? (March 2026)
Applications on the public cloud raise strong concerns about data protection. As an architect, I spend a meaningful part of my time ensuring the security of customers’ data in the cloud.
Bao introduces an innovative approach where data remains on local devices while the cloud provides encrypted storage for synchronization and peer exchange. Because cloud providers cannot access the data, the need for due diligence is reduced.
Any feedback is welcome.
The problem: every agent (Cline, Aider, Codex, Claude Code) has unrestricted access to your filesystem, shell, and network. When they process untrusted content — a cloned repo, a dependency README — they’re prompt injection vectors with full machine access. No existing tool evaluates what the agent actually does at the syscall level.
grith wraps any CLI agent without modification. OS-level interception captures every file open, network call, and process spawn, then runs it through 17 independent security filters in parallel across three phases (~15ms total). Composite score routes each call: auto-allow, auto-deny, or queue for async review. Most will auto approve - which eliminates approval fatigue.
Also does per-session cost tracking and audit trails as a side effect of intercepting everything.
I think the first step is standardizing HTTP 402 using traditional, familiar payment rails like Stripe, then we can move to things like on-chain or other rails later.
I am building https://stripe402.com to try to make it dead simple for those building APIs/resources to get paid per request through stripe without user's needing to sign up for accounts, get API keys, any of that normal painful workflow required today.
Check it out and feedback welcome!
I have worked with data for a while. I feel like our tools could be much better when it comes to "flow". I want an experience where you don't need to alt+tab to slack/images/another query. What if we put it all on a canvas? That's what Kavla is all about!
Since last month I've done a lot of improvements to the editor to make the "flow" better.
I've also read up on HMAC, Nonces and fun encryption stuff to create read only boards.
Here's one where I look at stack overflow survey for databases: https://app.kavla.dev/v/mqhg54o319doya4.67dbfee1ccd6caf638d3...
Snowflake users apparently make the most money!
https://community.learningequality.org/t/bringing-new-comput...
Also trying to recruit people to teach tech newbies how to build their own handheld video game consoles. Let me know if you might like to run a class where you live and i'll share my class materials.
https://community.arduboy.com/t/looking-for-instructors-to-t...
Basically OpenClaw but with investing dashboards for my portfolio, additional tools specifically for investing, and exploring an AI-Human collaboration on researching economics (check the 'community' tab).
The data models are all in markdown and Excel so that there's no lockin and you can manually edit positions, personalities, etc.
This comes from frustration around most investing tools basically scraping your personal data + forcing you to lock into subscriptions. I think it's now possible to just vibe code most of what one needs, aside form raw data subscriptions.
It's all open source, too: https://github.com/wgryc/athena-os
A lot of existing databases are storage first, with everything else built around them. I have been exploring what it looks like if the database is closer to the application runtime itself, where state is live, queryable, and easier to reason about directly.
One thing I am prototyping right now is database-native tests.
Basically: what if integration tests were a database primitive?
CREATE TEST test::insert { INSERT test::users [{ id: 99, name: "Ghost" }]; FROM test::users | FILTER id == 99 | ASSERT { name == "Ghost" }; };
So not a wrapper, not a framework, not an external test runner.
A real test object inside the database.
The idea is that you could run these before schema changes, and make stored procedures or other database logic much easier to test without leaving the database model.
Still early, but it feels like one of those things that should just exist, especially for databases built around live application state.