logoalt Hacker News

randomNumber7yesterday at 9:21 PM3 repliesview on HN

Local models just make no economic sense since the GPU will idle 99% of the time.


Replies

zozbot234yesterday at 10:00 PM

You have a GPU already (at least an iGPU and an NPU on most newer platforms) as part of your computer, might as well get some use out of it with local inference. And trying to do inference on a larger model with an undersized GPU will have you idling a lot less than 99% - but that still makes a lot of sense for most casual users who will only rarely need a genuine "Pro" class answer from AI. Doing that locally is way less hassle than paying for a subscription or messing with API spend.

amazingamazingyesterday at 11:50 PM

False on a team that’s distributed

twoodfinyesterday at 9:43 PM

[dead]