logoalt Hacker News

randomNumber7last Tuesday at 9:21 PM3 repliesview on HN

Local models just make no economic sense since the GPU will idle 99% of the time.


Replies

zozbot234last Tuesday at 10:00 PM

You have a GPU already (at least an iGPU and an NPU on most newer platforms) as part of your computer, might as well get some use out of it with local inference. And trying to do inference on a larger model with an undersized GPU will have you idling a lot less than 99% - but that still makes a lot of sense for most casual users who will only rarely need a genuine "Pro" class answer from AI. Doing that locally is way less hassle than paying for a subscription or messing with API spend.

amazingamazinglast Tuesday at 11:50 PM

False on a team that’s distributed

twoodfinlast Tuesday at 9:43 PM

[dead]