logoalt Hacker News

dpoloncsaklast Tuesday at 8:56 PM3 repliesview on HN

Yeah, one of my first projects one of my buddies asked "Why aren't you using [ChatGPT 4.0] nano? It's 99% the effectiveness with 10% the price."

I've been using the smaller models ever since. Nano/mini, flash, etc.


Replies

sixtyjlast Tuesday at 9:12 PM

Yup.

I have found out recently that Grok-4.1-fast has similar pricing (in cents) but 10x larger context window (2M tokens instead of ~128-200k of gpt-4-1-nano). And ~4% hallucination, lowest in blind tests in LLM arena.

show 1 reply
phainopepla2last Tuesday at 9:03 PM

I have been benchmarking many of my use cases, and the GPT Nano models have fallen completely flat one every single except for very short summaries. I would call them 25% effectiveness at best.

show 1 reply
walthamstowlast Tuesday at 9:04 PM

Flash Lite 2.5 is an unbelievably good model for the price