How much VRAM does it need? I haven't run a local model yet, but I did recently pick up a 16GB GPU, before they were discontinued.
Pick a decent quant (4-6KM) then use llama-fit-params and try it yourself to see if it's giving you what you need.
Should run just fine with CPU-MoE and mmap, but inference might be a bit slow if you have little RAM.
You can run 25-30b model easily if you use Q3 or Q4 quants and llama-server with a pretty long list of options.
If you have to ask then your GPU is too small.
With 16 GB you'll be only able to run a very compressed variant with noticable quality loss.
It's on the page: