logoalt Hacker News

the_harpia_ioyesterday at 6:47 AM3 repliesview on HN

Interesting approach for cost management, but one angle nobody seems to be discussing: the security implications.

When you fall back to a local model for coding, you lose whatever safety guardrails the hosted model has. Claude's hosted version has alignment training that catches some dangerous patterns (like generating code that exfiltrates env vars or writes overly permissive IAM policies). A local Llama or Mistral running raw won't have those same checks.

For side projects this probably doesn't matter. But if your Claude Code workflow involves writing auth flows, handling secrets, or touching production infra, the model you fall back to matters a lot. The generated code might be syntactically fine but miss security patterns that the larger model would catch.

Not saying don't do it - just worth being aware that "equivalent code generation" doesn't mean "equivalent security posture."


Replies

sReinwaldyesterday at 9:14 AM

Not saying the frontier models aren't smarter than the ones I can run on my two 4090s (they absolutely are) but I feel like you're exaggerating the security implications a bit.

We've seen some absolutely glaring security issues with vibe-coded apps / websites that did use Claude (most recently Moltbook).

No matter whether you're vibe coding with frontier models or local ones, you simply cannot rely on the model knowing what it is doing. Frankly, if you rely on the model's alignment training for writing secure authentication flows, you are doing it wrong. Claude Opus or Qwen3 Coder Next isn't responsible if you ship insecure code - you are.

show 1 reply
tossandthrowyesterday at 7:12 AM

Yes, models are aligned differently. But that is a quality of the model.

Obviously it must be assumed that the model one falls back on is good enough - including security alignment.

show 1 reply
Zardoz84yesterday at 6:55 AM

I would always prefer something local. By definition it's more secure, as you are not sending your code on the wire to a third party server, and hope that they comply with the "We will not train our models with your data".

show 1 reply