With local inference on pretty decent local models we have nowadays (Qwen-3.5 and better) it's not much of a concern anymore.
Sure it is - there's still an opportunity cost of spending tokens(time/energy) creating a library from scratch vs using a preexisting well understood API.
what percentage of people is using local models for anything serious? I reckon single digits if even that. And for a corporate work environment, probably close to 0.
Sure, if you've got a £5k laptop