Posted by jbredeche 8 hours ago
Pulled 30 real Twitter accounts from community-archive.org and built insurance-style risk models. The model scores: original content per day, vocabulary diversity, and bot resistance. Then prices the risk of the creator going silent.
Results: Total insurable annual output across 30 accounts = 449K. Monthly premiums = 1307. Five accounts flagged suspicious for high volume + low diversity. Zero confirmed bots (self-selected archive).
The interesting finding: accounts with bot-like patterns (high volume, low vocabulary diversity) naturally get priced OUT by the insurance model. You cannot insure what is not real content.
Feedback distance is not just a safety problem. It is an actuarial one. The shorter the distance between human and AI, the lower the insurance premium.
That's just straight up nonsense, no? How much cherry picking do you need?
I really hope this is a simulation example.
The way Clio works, "private" is just removing first person speech but leaving a summary of the data behind.
Even though the data is summarized, that still means that your ip is still stored by anthropic? For me it's actually a huge data security issue (that I only figured out now sigh).
So what is the point of me enabling privacy mode when it doesn't really do anything?
>from under 25 minutes to over 45 minutes.
If I get my raspberry pi to run a LLM task it'll run for over 6 hours. And groq will do it in 20 seconds.
It's a gibberish measurement in itself if you don't control for token speed (and quality of output).
The fact that there is no clear trend in lower percentiles makes this more suspect to me.
If you want to control for user base evolution given the growth they've seen, look at the percentiles by cohort.
I actually come away from this questioning the METR work on autonomy.
You can see the trend for other percentiles at the bottom of this, which they link to in the blog post https://cdn.sanity.io/files/4zrzovbb/website/5b4158dc1afb211...