Posted by david927 10 hours ago
Ask HN: What Are You Working On? (March 2026)
Pipeline so far has gone like this:
* Use the search engine's API to query a bunch of depravity
* Use qwen3.5 to label the search results and generate training data
* Try to use fasttext to create a fast model
* Get good results in theory but awful results in practice because it picks up weird features
* Yolo implement a small neural net using hand selected input features instead
* Train using fasttext training data
* Do a pretty good job
* for (;;) Apply the model to real a world link database and relabel positive findings with qwen to provide more training data
Currently this is where I'm at
Accuracy: 90.90%
True Positive: 1021
False Positive: 154
True Negative: 2816
False Negative: 230
Precision: 0.8689
Recall: 0.8161
F1: 0.8417
There's a lot of vague middle ground and many of the false positives are arguably just mislabeled.AIOs are a black hole - we dont know when they appear and whats in it. so i creates a tool thats starts with GSC data and enriches it via AIO data
works good and the major finding by now
the best AIOs you can get are ..... none.
doesn't matter if you are in it or not - as soon as they show up the CTR to tour web-property goes down massively ~60% to 70%
the CTR on the AIOs are ~0%
const app = new App("com.apple.finder")
and then query for elements: const window = app.$({role: "window"})
const someButton = window.$(/* another query */)
and then do stuff with it: someButton.press()
and you can bind everything to very specific shortcuts like "press and hold cmd, then scroll mouse wheel up"Targeted towards music producers and AI (there's one collection of snippets that starts an MCP server and exposes some basic functionality) in the beginning.
I started developing a city builder called Metropolis 1998 [1], but wanted to take the genre in new directions, building on top of what modern games have to offer:
- Watch what's happening inside buildings and design your own (optional)
- Change demand to a per-business level
- Bring the pixel art 3D render aesthetic back from the dead (e.g RollerCoaster Tycoon) [2]
I just updated my Steam page with some recent snapshots from my game. Im really happy with how the game is turning out!
[1] https://store.steampowered.com/app/2287430/Metropolis_1998/
[2] The art in my game is hand drawn though
> Both adults in a family will now own a car. This is required since there are not other transportation options, and sidewalks are optional.
Is this temporary or are you planning to release it like this? SimCity leaned into euclidean zoning (separate industrial/residential/commercial zones) and pocketable cars which needed no parking, and thus failed to properly showcase how ugly car-centric cities actually are. I’m sure they did it because it made for an easy gameplay loop/balancing but I’d hope we could come up with more realistic and interesting mechanics in 2026
Will you do a native Linux release, or has it been tested with Proton?
Also, just from watching the video and screenshots in the Steam page, it seems like a crazy amount of work. Are you doing everything by yourself?
Didn't realise you'd swapped to isometric, it's looking fabulous!
Did you roll your own engine, I know Godot has issues scaling past a certain number of simulations.
We’ve continued to get some paid customers and have exited beta last week, given everyone seemed to be quite satisfied and there hadn't been requests for changes, only some specific search providers.
Because of bots there isn’t a free trial easily available, but if you’re a human and you’d like to try it for a couple of days for free, reach out with your account number and we’ll set that up!
Thanks.
P.S.: Because people have asked before, our tech stack is intentionally very "boring" (as in, it generates and serves the HTML + bits of JS to enhance settings and such — search can be done without JS), using Deno in the backend (for easier TypeScript), PostgreSQL for the DB, and Docker for easier deploying.
Making improvements on this tetris meets block puzzle game
Also working on a language for embedded bare-metal devices with built-in cooperative multitasking.
A lot of embedded projects introduce an RTOS and then end up inheriting the complexity that comes with it. The idea here is to keep the mental model simple: every `[]` block runs independently and automatically yields after each logical line of code.
There is also an event/messaging system:
- Blocks can be triggered by events: `[>event params ...]`
- Blocks can wait for events internally
- Events can also be injected from interrupts
This makes it easy to model embedded systems as independent state machines while still monitoring device state.
Right now it’s mostly an interpreter written in Rust, but it can also emit C code. I’m still experimenting with syntax.
Example:
module WaterTank {
type Direction = UP|DOWN
let direction = UP
let current = 0
[>open_valve direction |> direction]
[>update level |> current]
[
for 0..30 |> iteration {
when direction {
UP -> !update level=current + 1 |> min(100)
DOWN -> !update level=current - 1 |> max(0)
} ~
%'{iteration} {current}'
}
]
[>update level |> when {
0..10 -> %'shallow'
11..15 -> %'good'
16.. -> %'too much!' then !open_valve direction=DOWN
}
]
}I have been working on it as side project for over two years and now, with funding from the EU for the next 2.5 years, I hope I can make of it a real product for everyone to use that can compete with the likes of Excel and Googl;e Sheets.
I can oly say, I am overly, off the Moon excited
edit: nm, rtfm, it was on the landing page: Horizon Europe programme
NLnet is just amazing and can keep you going if you are a student or have some extra sources of income
HORIZON is a huge grant but fairly hard to obtain. Generally related to reasearch grants in academia
As you see there is a huge component of sheer luck
[1]: https://nlnet.nl/project/IronCalc/ [2]: https://nextgraph.org/ [3]: https://elfaconsortium.eu/
The content is hand picked from tiktok, Instagram, Facebook, Reddit and other AI generating platforms.
Honestly I don't know where I'm going with this, but I felt the urge to create it, so here it is.
I learned how to optimize serving assets on CloudFlare.
Feedback welcome.
EDIT: Hm, I switched tab, away to write this comment, now that I switched back, it showed me that I clicked correctly. So it seems, that sometimes it just has huge delay in accepting my choice?
Edit: I don't see slow traces in Sentry. No idea what caused this. Also, voting goes through redis and the dB load is low. Weird. I probably have to add gunicorn workers.
Edit2: Bumped gunicorn workers from 2 to 4. Should be fine now, under the current load. Again, thank you for reporting!
I dunno if/how this could be taught, but I feel like half the battle is critical thinking with an adversarial mindset towards media -- who would make this, why would they want to show me, do I see anything that makes this impossible, is it worth engaging with in the first place, can I fact check this.
I'm trying to gamify the training to make the experience more appealing.
I store a "proof URL" on the backend, but I don't know if it makes sense to serve it to the end user. Also, a Reddit discussion is not necessarily a proof one wants. A fingerprint would be better, but not all images are generated with Google. That's another problem to be solved.
It's SFW and localized to the most popular languages.