logoalt Hacker News

Local AI needs to be the norm

1819 pointsby cylolast Sunday at 5:19 PM726 commentsview on HN

Comments

moveax3yesterday at 10:14 AM

[dead]

ElenaDaibunnyyesterday at 7:11 AM

[dead]

throwaway613746last Sunday at 9:57 PM

[dead]

debpalashlast Sunday at 9:41 PM

[dead]

ki_sum_aiyesterday at 3:41 PM

[flagged]

system2yesterday at 7:20 AM

[flagged]

barrkellast Sunday at 8:19 PM

Local models are extraordinarily expensive if you're not maximizing throughput, and you're not going to be maximizing it.

Local models need to be resident in expensive RAM, the kind that has fat pipes to compute. And if you have a local app, how do you take a dependency on whatever random model is installed? Does it support your tool calling complexity? Does it have multimodal input? Does it support system messages in the middle of the conversation or not? Is it dumb enough to need reminders all the time?

Spend enough time building against local models and you'll see they're jagged in performance. You need to tune context size, trade off system message complexity with progressive disclosure. You simply can't rely on intelligence. A bunch of work goes into the harness.

Meanwhile, third party inference is getting the benefits of scale. You only need to rent a timeslice of memory and compute. It's consistent and everybody gets the same experience. And yes, it needs paying for, but the economics are just better.

show 3 replies