prismML provides a llama.cpp fork which is compatible with the 1 bit models:
https://github.com/PrismML-Eng/llama.cpp
After fails with Ollama and main llama.cpp the fork worked on my M5 MBA.
Edit: Typos