logoalt Hacker News

gnatmanyesterday at 8:54 PM3 repliesview on HN

LLMs sure do love to burn tokens. It’s like a high schooler trying to meet the minimum word length on a take home essay.


Replies

estimator7292yesterday at 8:59 PM

I've always wondered about that. LLM providers could easily decimate the cost of inference if they got the models to just stop emitting so much hot air. I don't understand why OpenAI wants to pay 3x the cost to generate a response when two thirds of those tokens are meaningless noise.

show 4 replies
sambaumannyesterday at 9:00 PM

I feel like this has gotten much worse since they were introduced. I guess they're optimizing for verbosity in training so they can charge for more tokens. It makes chat interfaces much harder to use IMO.

I tried using a custom instruction in chatGPT to make responses shorter but I found the output was often nonsensical when I did this

show 1 reply
zwaragyesterday at 9:24 PM

well, they probably have quite a lot of text from high schoolers trying to meet the minimum word length on a take home essay in the training data