logoalt Hacker News

Can I run AI locally?

654 pointsby ricardbejaranotoday at 12:46 PM187 commentsview on HN

Comments

golem14today at 6:02 PM

Has anyone actually built anything with this tool?

The website says that code export is not working yet.

That’s a very strange way to advertise yourself.

adithyassekhartoday at 4:21 PM

This just reminded me of this https://www.systemrequirementslab.com/cyri.

Not sure if it still works.

havaloctoday at 4:34 PM

Missing the A18 Neo! :)

debatem1today at 4:32 PM

For me the "can run" filter says "S/A/B" but lists S, A, B, and C and the "tight fit" filter says "C/D" but lists F.

Just FYI.

lagrange77today at 7:05 PM

Finally! I've been waiting for something like this.

ThrowawayTestrtoday at 8:46 PM

For image generation or even video generation, local models are totally feasible. I can generate a 5 second clip with wan 2.2 in about 30 minutes on my 3060 12G. Plus, I have full control on the loras used.

arjietoday at 4:44 PM

Cool website. The one that I'd really like to see there is the RTX 6000 Pro Blackwell 96 GB, though.

ameliustoday at 5:32 PM

What is this S/A/B/C/etc. ranking? Is anyone else using it?

show 2 replies
tencentshilltoday at 6:44 PM

Missing laptop versions of all these chips.

jrmgtoday at 5:09 PM

Is there a reliable guide somewhere to setting up local AI for coding (please don’t say ‘just Google it’ - that just results in a morass of AI slop/SEO pages with out of date, non-self-consistent, incorrect or impossible instructions).

I’d like to be able to use a local model (which one?) to power Copilot in vscode, and run coding agent(s) (not general purpose OpenClaw-like agents) on my M2 MacBook. I know it’ll be slow.

I suspect this is actually fairly easy to set up - if you know how.

show 3 replies
ameliustoday at 5:31 PM

Why isn't there some kind of benchmark score in the list?

ryandraketoday at 6:19 PM

Missing RTX A4000 20GB from the GPU list.

S4phyretoday at 4:16 PM

Oh how cool. Always wanted to have a tool like this.

bheadmastertoday at 6:42 PM

Missing 5060 Ti 16GB

brcmthrowawaytoday at 5:50 PM

If anyone hasn't tried Qwen3.5 on Apple Silicon, I highly suggest you to! Claude level performance on local hardware. If the Qwen team didn't get fired, I would be bullish on Local LLM.

ipunchghoststoday at 7:41 PM

What is S? Also, NVIDIA RTX 4500 Ada is missing.

varispeedtoday at 5:16 PM

Does it make any sense? I tried few models at 128GB and it's all pretty much rubbish. Yes they do give coherent answers, sometimes they are even correct, but most of the time it is just plain wrong. I find it massive waste of time.

show 1 reply
g_br_ltoday at 4:31 PM

could you add raspi to the list to see which ridiculously small models it can run?

metalliqaztoday at 4:33 PM

Hugging Face can already do this for you (with much more up-to-date list of available models). Also LM Studio. However they don't attempt to estimate tok/sec, so that's a cool feature. However I don't really trust those numbers that much because it is not incorporating information about the CPU, etc. True GPU offload isn't often possible on consumer PC hardware. Also there are different quants available that make a big difference.

charcircuittoday at 4:32 PM

On mobile it does not show the name of the model in favor of the other stats.

tristortoday at 6:39 PM

This does not seem accurate based on my recently received M5 Max 128GB MBP. I think there's some estimates/guesswork involved, and it's also discounting that you can move the memory divider on Unified Memory devices like Apple Silicon and AMD AI Max 395+.

polyterativetoday at 6:36 PM

awesome, needed this

kylehotchkisstoday at 5:23 PM

My Mac mini rocks qwen2.5 14b at a lightning fast 11/tokens a second. Which is actually good enough for the long term data processing I make it spend all day doing. It doesn’t lock up the machine or prevent its primary purpose as webserver from being fulfilled.

nilslindemanntoday at 5:29 PM

1. More title attributes please ("S 16 A 7 B 7 C 0 D 4 F 34", huh?)

2. Add a 150% size bonus to your site.

Otherwise, cool site, bookmarked.

tkfosstoday at 6:57 PM

Nice UI, but crap data, probably llm generated.

JulianPembroketoday at 6:51 PM

[flagged]

Felixbottoday at 4:52 PM

[flagged]

show 1 reply
Yanko_11today at 7:31 PM

[dead]

JulianPembroketoday at 6:52 PM

[flagged]

show 1 reply
aplomb1026today at 5:31 PM

[dead]

A7OMtoday at 6:51 PM

[flagged]

A7OMtoday at 6:50 PM

[flagged]

prokajevotoday at 6:26 PM

[dead]

uncSofttoday at 4:51 PM

[dead]

unfirehosetoday at 5:11 PM

if you do, would you still want to collect data in a single pane of glass? see my open source repo for aggregating harness data from multiple machine learning model harnesses & models into a single place to discover what you are working on & spending time & money. there is plans for a scrobble feature like last.fm but for agent research & code development & execution.

https://github.com/russellballestrini/unfirehose-nextjs-logg...

thanks, I'll check for comments, feel free to fork but if you want to contribute you'll have to find me off of github, I develop privately on my own self hosted gitlab server. good luck & God bless.