DwarfStar4 is a small LLM inference runtime that can run DeepSeek 4. The blog post implies that it currently requires 96GB of VRAM.
For others who are lacking context :-)
That's the flash version not the full model and only at Q2-3~ so while impressive it's still quite different from the full model.
> The blog post implies that it currently requires 96GB of VRAM.
Has anyone tested what happens if you try and run this on lower-RAM Macs? It might work and just be a bit slower as it falls back on fetching model layers from storage.
Thanks. How is DwarfStar4 different from llama.cpp?
I knew Death Stranding 3 wasn't out yet!
>The blog post implies that it currently requires 96GB of VRAM.
From the Github page it seems it only supports Apple and DGX Spark. I have 128 GB of RAM and a 3090 but it probably won't work.
Thanks. Outside of LLM circles, DS4 is usually a video game controller.