btw, don't bother trying to buy a bunch of Mac boxes to run LLMs in parallel because it won't be any faster than a single box.
is everyone just waiting for teh DGX Spark? Are they really going to ban local inference?
is everyone just waiting for teh DGX Spark? Are they really going to ban local inference?