Remix.run Logo
zettabomb 5 days ago

llama.cpp has built-in support for doing this, and it works quite well. Lots of people running LLMs on limited local hardware use it.

EnPissant 4 days ago | parent [-]

llama.cpp has support for running some of or all of the layers on the CPU. It does not swap them into the GPU as needed.