▲ | refulgentis 5 days ago | |||||||||||||
Nah, llama.cpp is stable. llama.cpp also got GPT-OSS early, like Ollama. There's a lot of extremely subtle politics going on in the link. Suffice it to say, as a commercial entity, there's a very clever way to put your thumb on the scale of what works and what doesn't without it being obvious to anyone involved, even the thumb. | ||||||||||||||
▲ | hodgehog11 5 days ago | parent | next [-] | |||||||||||||
Stable for a power user, or stable for everyone? I don't have links on hand, but I could swear there have been instances where certain models rolled back support during llama.cpp development, and this was recent. Also llama.cpp adds features and support on a near-daily basis, how can this be LTS? Don't get me wrong, llama.cpp is an amazing tool. But it's development is nowhere near as cautious as something like the Linux kernel, so there is room there for a more stable alternative. Not saying Ollama will do this, but llama.cpp won't be everything to everyone. | ||||||||||||||
| ||||||||||||||
▲ | mhitza 5 days ago | parent | prev [-] | |||||||||||||
llama.cpp still doesn't support gpt-oss tool calling. https://github.com/ggml-org/llama.cpp/pull/15158 (among other similar PRs) But I also couldn't get vllm, or transformers serve, or ollama (400 response on /v1/chat/completions) working today with gpt-oss. OpenAI's cookbooks aren't really copy paste instructions. They probably tested on a single platform with preinstalled python packages which they forgot to mention :)) | ||||||||||||||
|