logoalt Hacker News

WAyesterday at 7:07 AM1 replyview on HN

Why though? The context window is 1 millions token max so far. That is what, a few MB of text? Sounds like I should be able to run claw on a raspberry pi.


Replies

tjchearyesterday at 2:08 PM

If you’re using it with a local model then you need a lot of GPU memory to load up the model. Unified memory is great here since you can basically use almost all the RAM to load the model.