logoalt Hacker News

evo_9today at 2:31 PM7 repliesview on HN

Every AI subscription is a ticking time bomb for the frontier provider; within a few years we will be running local models as good as today’s frontier models with almost no cost burden. The floor will fall out of the enterprise market for all the frontier companies.


Replies

crazygringotoday at 3:20 PM

> within a few years we will be running local models as good as today’s frontier models with almost no cost burden

Based on what? The RAM requirements alone are extraordinary.

No, running large models on shared, dedicated hosted hardware at full utilization is going to be vastly more cost-efficient for the foreseeable future.

show 1 reply
adamgordonbelltoday at 2:41 PM

Or put another way, the frontier models are very quickly deprecating assets, because of the competition in the market.

They have to keep getting better to stay ahead of each other and open weight.

Which means it's the opposite of a timebomb, the article has it completely backwards, tokens at current level of reasoning will continue to get cheaper.

I'm not sure 'local' will be the end state, as hardware needs are high. But certainly competitive forces tend to push profit margins toward zero.

Extended discussion on this topic:

https://corecursive.com/the-pre-training-wall-and-the-treadm...

show 1 reply
vb-8448today at 3:23 PM

> within a few years we will be running local models as good as today’s frontier

Unless there isn't some important breakthrough in hw production or in models architecture, it's quite the opposite: bigger, more expensive and more energy-intensive hw is needed today compared to 1 or 2 years ago.

show 1 reply
wolttamtoday at 3:21 PM

There's still going to be plenty of use-case and demand for frontier models running across hundreds or thousands of GPUs. It's just not going to be in the current shape - certainly not accessed by the general public for rote business tasks.

nijavetoday at 3:01 PM

>within a few years

Eventually, we'll see. Frontier models still need some pretty serious hardware which will slowly come down in cost. Smaller models are becoming more capable, which will presumably continue to improve.

I think there's still a pretty big gap, though. Claude estimates Opus 4.6 and GLM-5 need about 1.5Ti VRAM. It puts gpt-5.5 around 3-6Ti of VRAM.

That's 8x Nvidia H200 @ ~$30k USD each. Still need some big efficiency improvements and big hardware cost reduction.

show 1 reply
YesBoxtoday at 3:24 PM

You'd have a point if Cloud ^tm didnt take off into a multi billion dollar industry.

guesswho_today at 2:42 PM

[dead]