logoalt Hacker News

AussieWog93yesterday at 9:56 PM0 repliesview on HN

I've tried these small models and they're nowhere near as good as Claude or GPT-5.

The new ones running on a 16GB M1 are maybe GPT-4 level (with decent performance to be fair).

I wonder if it's possible to make some hyper-overturned model that, say, does nothing but program in Python get SOTA-ish performance in that narrow task.