Remix.run Logo
wing-_-nuts 9 hours ago

Llama.cpp is one of those projects that I want to install, but I always just wind up installing kobold.cpp because it's simply miles better with UX.

lolinder 7 hours ago | parent | next [-]

Llama.cpp forms the base for both Ollama and Kobold.cpp and probably a bunch of others I'm not familiar with. It's less a question of whether you want to use llama.cpp or one of the others and more of a question of whether you benefit from using one of the wrappers.

I can imagine some use cases where you'd really want to use llama.cpp directly, and there are of course always people who will argue that all wrappers are bad wrappers, but for myself I like the combination of ease of use and flexibility offered by Ollama. I wrap it in Open WebUI for a GUI, but I also have some apps that reach out to Ollama directly.

phillipcarter 7 hours ago | parent | prev | next [-]

I just use ollama. It works on my mac and windows machine and it's super simple to install + run most open models. And you can have another tool just shell out to it if you want more than than the CLI.

syntaxing 8 hours ago | parent | prev | next [-]

Llama cpp is more of a backend software. Most front end software like kobold/open webui uses it

tempest_ 8 hours ago | parent [-]

I found it only took me ~20 minutes to get Open-WebUI and Ollama going on my machine locally. I don't really know what is happening under the hood but from 0 to chat interface was definitely not too hard.

cwillu 8 hours ago | parent | prev [-]

“koboldcpp forked from ggerganov/llama.cpp”