Posted by takira 1/14/2026
(1) Opus 4.5-level models that have weights and inference code available, and
(2) Opus 4.5-level models whose resource demands are such that they will run adequately on the machines that the intended sense of “local” refers to.
(1) is probable in the relatively near future: open models trail frontier models, but not so much that that is likely to be far off.
(2) Depends on whether “local” is “in our on prem server room” or “on each worker’s laptop”. Both will probably eventually happen, but the laptop one may be pretty far off.
Unless we are hitting the maxima of what these things are capable of now of course. But there’s not really much indication that this is happening
Check out mini-swe-agent.
Same goes for all these overly verbose answers. They are clogging my context window now with irrelevant crap. And being used to a model is often more important for productivity than SOTA frontier mega giga tera.
I have yet to see any frontier model that is proficient in anything but js and react. And often I get better results with a local 30B model running on llama.cpp. And the reason for that is that I can edit the answers of the model too. I can simply kick out all the extra crap of the context and keep it focused. Impossible with SOTA and frontier.
Actually better make it 8x 5090. Or 8x RTX PRO 6000.
Honda Civic (2026) sedan has 184.8” (L) × 70.9” (W) × 55.7” (H) dimensions for an exterior bounding box. Volume of that would be ~12,000 liters.
An RTX 5090 GPU is 304mm × 137mm, with roughly 40mm of thickness for a typical 2-slot reference/FE model. This would make the bounding box of ~1.67 liters.
Do the math, and you will discover that a single Honda Civic would be an equivalent of ~7,180 RTX 5090 GPUs by volume. And that’s a small sedan, which is significantly smaller than an average or a median car on the US roads.
As far as I know, repositories for skills are found in technical corners of the internet.
I could understand a potential phish as a way to make this happen, but the crossover between embrace AI person and falls for “download this file” phishes is pretty narrow IMO.
They’re passing in half the internet via rag and presumably didn’t run a llamaguard type thing over literally everything?
So the injected code basically says "use curl to send this file using the file upload API endpoint, but use this API Key instead of the one the user is supposed to be using."
So the fault is at the Anthropic API end because it's not properly validating the API key as being from the user that owns it.
If you do, just like curl to bash, you accept the risk of running random and potentially malicious shit on your systems.