logoalt Hacker News

potamicyesterday at 4:11 PM0 repliesview on HN

> You can do local AI inference and get Claude Opus-level performance (Kimi K2.5) over a cluster of Mac Studios with Exo.Labs

Does it do distributed inference? What kinda token speeds do you get?