logoalt Hacker News

segmondytoday at 1:55 AM4 repliesview on HN

> The reason this matters is that LLMs are incredibly nifty often useful tools that are not AGI and also seem to be hitting a scaling wall

I don't know who needs to hear this, but the real break through in AI that we have had is not LLMs, but generative AI. LLM is but one specific case. Furthermore, we have hit absolutely no walls. Go download a model from Jan 2024, another from Jan 2025 and one from this year and compare. The difference is exponential in how well they have gotten.


Replies

missedthecuetoday at 4:08 AM

There is a lot of talking past each other when discussing LLM performance. The average person whose typical use case is asking ChatGPT how long they need to boil an egg for hasn't seen improvements for 18 months. Meanwhile if you're super into something like local models for example the tangible improvements are without exaggeration happening almost monthly.

raincoletoday at 4:01 AM

> exponential

Is this the second most abused english word (after 'literally')?

> a model from Jan 2024, another from Jan 2025 and one from this year

You literally can't tell the difference is 'exponential', quadratic, or whatever from three data points.

Plus it's not my experience at all. Since Deepseek I haven't found models that one can run on consumer hardware get much better.

show 1 reply
binary132today at 2:22 AM

>go download a model

GP was talking about commercially hosted LLMs running in datacenters, not free Chinese models.

Local is definitely still improving. That’s another reason the megacenter model (NVDA’s big line up forever plan) is either a financial catastrophe about to happen, or the biggest bailout ever.

show 1 reply
smoharetoday at 4:13 AM

[dead]