To me the article isn’t convincing. There are some interesting points raised. Like “the llm sharing memories between a team of developers”.
I mostly use llms individually, so this is a real blind spot for me. But I might convinced to share a corpus of memories between developers if it ever becomes practical.
But for now aren’t context windows still sometimes smaller than the task at hand… so those “llm memories” take the form of literal documentation???
The closer llm memories get to native format, i.e: stored tokenized content, don’t we lose compatibility between setups anyhow. What if you’re using fp16, and I’m using nvfp4?