logoalt Hacker News

digitaltreesyesterday at 1:00 AM3 repliesview on HN

I wonder if it really needs to be worse. I am playing with the idea of fine tuning a model on my exact stack and coding patterns. I suspect I could get better performance by training “taste” into a model rather than breadth.


Replies

epicureanidealyesterday at 4:20 AM

I also wonder about JS only, Python only, etc models.

Maybe the future is a selection of local, specific stack trained models?

show 2 replies
jimbokunyesterday at 5:58 PM

That approach has its advantages, but sometimes I want to generate code for a language or kind of project I’m not experienced with using the accepted best practices.

andy_pppyesterday at 4:45 AM

Fine tuning these models (at least with PPO or equivalent) requires even more VRAM than inference does, potentially 2-3 times more.

show 1 reply