| ▲ | WA 16 hours ago | |
Why though? The context window is 1 millions token max so far. That is what, a few MB of text? Sounds like I should be able to run claw on a raspberry pi. | ||
| ▲ | tjchear 9 hours ago | parent [-] | |
If you’re using it with a local model then you need a lot of GPU memory to load up the model. Unified memory is great here since you can basically use almost all the RAM to load the model. | ||