▲ | wing-_-nuts 10 months ago | ||||||||||||||||
Llama.cpp is one of those projects that I want to install, but I always just wind up installing kobold.cpp because it's simply miles better with UX. | |||||||||||||||||
▲ | lolinder 10 months ago | parent | next [-] | ||||||||||||||||
Llama.cpp forms the base for both Ollama and Kobold.cpp and probably a bunch of others I'm not familiar with. It's less a question of whether you want to use llama.cpp or one of the others and more of a question of whether you benefit from using one of the wrappers. I can imagine some use cases where you'd really want to use llama.cpp directly, and there are of course always people who will argue that all wrappers are bad wrappers, but for myself I like the combination of ease of use and flexibility offered by Ollama. I wrap it in Open WebUI for a GUI, but I also have some apps that reach out to Ollama directly. | |||||||||||||||||
| |||||||||||||||||
▲ | phillipcarter 10 months ago | parent | prev | next [-] | ||||||||||||||||
I just use ollama. It works on my mac and windows machine and it's super simple to install + run most open models. And you can have another tool just shell out to it if you want more than than the CLI. | |||||||||||||||||
▲ | syntaxing 10 months ago | parent | prev | next [-] | ||||||||||||||||
Llama cpp is more of a backend software. Most front end software like kobold/open webui uses it | |||||||||||||||||
| |||||||||||||||||
▲ | xyc 10 months ago | parent | prev | next [-] | ||||||||||||||||
If anyone on macOS wants to use llama.cpp with ease, check out https://recurse.chat/. Supports importing ChatGPT history & continue chats offline using llama.cpp. Built this so I can use local AI as a daily driver. | |||||||||||||||||
▲ | cwillu 10 months ago | parent | prev [-] | ||||||||||||||||
“koboldcpp forked from ggerganov/llama.cpp” |