logoalt Hacker News

chr15mtoday at 1:54 PM1 replyview on HN

Is this something that will show up in Ollama any time soon to increase context size of local models?


Replies

zozbot234today at 2:56 PM

KV quantization has long been available in llama.cpp