Top
Best
New

Posted by tosh 2 days ago

Caveman: Why use many token when few token do trick(github.com)
876 points | 360 commentspage 9
bhwoo48 2 days ago|
I was actually worried about high token costs while building my own project (infra bundle generator), and this gave me a good laugh + some solid ideas. 75% reduction is insane. Starred
saidnooneever 2 days ago||
LOL it actually reads how humans reply the name is too clever :').

Not sure how effective it will be to dirve down costs, but honestly it will make my day not to have to read through entire essays about some trivial solution.

tldr; Claude skill, short output, ++good.

yesthisiswes 2 days ago||
Why use lot word when few word do fine.
kukakike 2 days ago||
This is exactly what annoys me most. English is not suitable for computer-human interaction. We should create new programming and query languages for that. We are again in cobol mindset. LLM are not humans and we should stop talking to them as if they are.
zozbot234 2 days ago|
Grug says Chinese more suitable, only few runes in word, each take single token. Is great.
throwatdem12311 2 days ago||
Ok but when the model is responding to you isn’t the text it’s generating also part of the context it’s using to generate the next token as it goes? Wouldn’t this just make the answers…dumb?
sillyboi 2 days ago||
Oh, another new trend! I love these home-brewed LLM optimizers. They start with XML, then JSON, then something totally different. The author conveniently ignores the system prompt that works for everything, and the extra inference work. So, it's only worth using if you just like this response style, just my two cents. All the real optimizations happen during model training and in the infrastructure itself.
thorfinnn 2 days ago||
kevin would be proud
Robdel12 2 days ago||
I didn’t comment on this when I saw it on threads/twitter. But it made it to HN, surprisingly.

I have a feeling these same people will complain “my model is so dumb!”. There’s a reason why Claude had that “you’re absolutely right!” for a while. Or codex’s “you’re right to push on this”.

We’re basically just gaslighting GPUs. That wall of text is kinda needed right now.

hybrid_study 2 days ago||
Mongo! No caveman
jongjong 1 day ago|
Me think this good idea. Regular language unnecessary complex. Distract meaning. Me wish everyone always talk this way. No hidden spin manipulate emotion. Information only. Complexity stupid.
More comments...