logoalt Hacker News

Caveman: Why use many token when few token do trick

587 pointsby toshtoday at 8:56 AM282 commentsview on HN

Comments

anshumankmrtoday at 3:30 PM

Though I do use Claude Code, is it possible to get this for Github Copilot too?

show 1 reply
rsynnotttoday at 6:26 PM

I mean, I assume you run into the same problem as Kevin in the office; that sort of faux-simple speech is actually very ambiguous.

(Though, I wonder has anyone tried Newspeak.)

fzeindltoday at 12:00 PM

I tried this with early ChatGPT. Asked it to answer telegram style with as few tokens as possible. It is also interesting to ask it for jokes in this mode.

show 1 reply
fnytoday at 12:29 PM

Are there any good studies or benchmarks about compressed output and performance? I see a lot of arguing in the comments but little evidence.

doe88today at 10:51 AM

> If caveman save you mass token, mass money — leave mass star.

Mass fun. Starred.

yesthisiswestoday at 4:50 PM

Why use lot word when few word do fine.

contingenciestoday at 7:22 PM

Better: use classical Chinese.

owenthejumpertoday at 11:57 AM

What is that binary file caveman.skill that I cannot read easily, and is it going to hack my computer.

adam_patarinotoday at 12:33 PM

Or you could use a local model where you’re not constrained by tokens. Like rig.ai

show 1 reply
throwatdem12311today at 4:25 PM

Ok but when the model is responding to you isn’t the text it’s generating also part of the context it’s using to generate the next token as it goes? Wouldn’t this just make the answers…dumb?

Suractoday at 7:45 PM

me like that

cadamsdotcomtoday at 10:40 AM

Caveman need invent chalk and chart make argument backed by more than good feel.

bitwizetoday at 1:13 PM

grug have to use big brains' thinking machine these days, or no shiny rock. complexity demon love thinking machine. grug appreciate attempt to make thinking machine talk on grug level, maybe it help keep complexity demon away.

us321today at 5:59 PM

I like

kukakiketoday at 11:59 AM

This is exactly what annoys me most. English is not suitable for computer-human interaction. We should create new programming and query languages for that. We are again in cobol mindset. LLM are not humans and we should stop talking to them as if they are.

show 1 reply
saidnooneevertoday at 10:25 AM

LOL it actually reads how humans reply the name is too clever :').

Not sure how effective it will be to dirve down costs, but honestly it will make my day not to have to read through entire essays about some trivial solution.

tldr; Claude skill, short output, ++good.

xgulfietoday at 1:58 PM

Funny how people are so critical of this and yet fawn over TOON

xpetoday at 1:37 PM

Unfrozen caveman lawyer here. Did "talk like caveman" make code more bad? Make unsubst... (AARG) FAKE claims? You deserve compen... AAARG ... money. AMA.

bogtogtoday at 10:31 AM

I'd be curious if there were some measurements of the final effects, since presumably models wont <think> in caveman speak nor code like that

isuckatcodingtoday at 3:55 PM

Oh come on now one referenced this scene from the office??

https://youtu.be/_K-L9uhsBLM?si=ePiGrFd546jFYZd8

sillyboitoday at 12:22 PM

Oh, another new trend! I love these home-brewed LLM optimizers. They start with XML, then JSON, then something totally different. The author conveniently ignores the system prompt that works for everything, and the extra inference work. So, it's only worth using if you just like this response style, just my two cents. All the real optimizations happen during model training and in the infrastructure itself.

Robdel12today at 12:39 PM

I didn’t comment on this when I saw it on threads/twitter. But it made it to HN, surprisingly.

I have a feeling these same people will complain “my model is so dumb!”. There’s a reason why Claude had that “you’re absolutely right!” for a while. Or codex’s “you’re right to push on this”.

We’re basically just gaslighting GPUs. That wall of text is kinda needed right now.

thorfinnntoday at 4:43 PM

kevin would be proud

vova_hn2today at 11:21 AM

I don't know about token savings, but I find the "caveman style" much easier to read and understand than typical LLM-slop.

hybrid_studytoday at 12:24 PM

Mongo! No caveman

bhwoo48today at 10:21 AM

I was actually worried about high token costs while building my own project (infra bundle generator), and this gave me a good laugh + some solid ideas. 75% reduction is insane. Starred

DonHopkinstoday at 12:26 PM

Deep digging cave man code reviews are Tha Shiznit:

https://www.youtube.com/watch?v=KYqovHffGE8

setnonetoday at 11:02 AM

caveman multilingo? how sound?

meidad_gtoday at 3:35 PM

[dead]

Adam_ciphertoday at 3:58 PM

[dead]

globalchatadstoday at 6:11 PM

[dead]

tatrionstoday at 4:18 PM

[dead]