18gb RAM it is a bit tight
with 32gb RAM:
qwen3-coder and glm 4.7 flash are both impressive 30b parameter models
not on the level of gpt 5.2 codex but small enough to run locally (w/ 32gb RAM 4bit quantized) and quite capable
but it is just a matter of time I think until we get quite capable coding models that will be able to run with less RAM
ahem ... cortex.build
Current test version runs in 8GB @ 60tks. Lmk if you want to join our early tester group!