Top
Best
New

Posted by lsdmtme 18 hours ago

Anthropic downgraded cache TTL on March 6th(github.com)
449 points | 341 commentspage 4
computerex 3 hours ago|
Good job anthropic. You had a clear lead with all devs singing the praises of Opus. Way to lose all that by Enshittifying the experience.
sscaryterry 15 hours ago||
Anthropic is leaving so much evidence around… proving damages and a pattern is becoming trivial
taffydavid 13 hours ago||
This is the same shit openAI used to do last year, quietly downgrading their offerings while hyping the next big thing. I thought Anthropic were different but it seems they're playing the exact same long con with Mythos.

They can't really revolutionize AI again so they make the product worse and worse and then offer you a "better" one

coffinbirth 15 hours ago||
Am I the only one who sees striking parallels between being a Claude Code customer and Cuckoldry (as in biology)?

I mean, you are investing a lot (infrastructure and capital) into something that is essentially not yours. You claim credit for the offspring (the solution) simply because it resides in your workspace. You accept foreign code to make your project appear more successful and populated than you could manage alone. Your over-reliance on a surrogate for the heavy lifting leads to the loss of your own survival skills (coding and debugging). Last but not least, you handle the grunt work of territory defense (clients and environments) while the AI performs the actual act of creation (Displaced Agency).

the_gipsy 14 hours ago||
What you're looking for is "vendor lock-in".
PunchyHamster 14 hours ago||
No, but it's very funny, I'm gonna call people that offshore their thinking to LLM "AI cucks" now
simianwords 15 hours ago||
There’s a case for intelligent caching: coarse grained 1h and 5min type TTls are not optimal.
PunchyHamster 14 hours ago|
Caching LLM is not like caching normal content; the longer it is the more beneficial it is and it only stops being worth when user stops current session.

So you'd need some adaptive algorithm to decide when to keep caching and when to purge it whole, possibly on client side, but if you give client the control, people will make it use most cache possible just to chase diminishing returns. So fine grained control here isn't all that easy; other possible option is just to have cache size per account and then intelligently purge it instead of relying just on TTL

cyanydeez 13 hours ago|||
keep in mind, efficient KV caching needs to be next to the GPU, so you sls need you HA to keep routing the user to the same hardware.

the hardware VM model is almost identical. Each session can go anywhere to start but a live session cant just be routed anywhere without penalty.

yobid20 5 hours ago||
i thought it was always 5 minutes? ive been telling people 5 minutes for months so i dont think this is anything new?
mrdw 11 hours ago||
I noticed another limitation: "An image in the conversation exceeds the dimension limit for many-image requests (2000px). Start a new session with fewer images."

So I can't continue my claude code session I started yesterday.

sunnybeetroot 3 hours ago||
Double tap ESC and revert the conversation.
beering 4 hours ago||
makes sense, “a picture is worth a thousand tokens” as they say. They probably lowered the limit due to capacity issues.
ares623 14 hours ago||
AGI finding bugs again. Actual Guys/Gals Instead.
idrdex 4 hours ago|
[dead]
More comments...