logoalt Hacker News

tuananhyesterday at 6:30 AM1 replyview on HN

yeah that goes without saying. how can openweight, quantized version beat SOTA :)


Replies

array_key_firstyesterday at 4:38 PM

Well it depends on the task. For agentic coding, more is more, but for tasks that normal consumers use them for there really is a ceiling. OCR, text to speech, that type of thing doesn't really improve when going to a SOTA model, so you'd just be wasting your money. I think local LLMs have more value than software engineers give them credit for.

show 1 reply