▲ | jnmandal 7 days ago | |||||||||||||
Honestly, I think it just depends. A few hours ago I wrote I would never want it for a production setting but actually if I was standing something up myself and I could just download headless ollama and know it would work. Hey, that would also be fine most likely. Maybe later on I'd revisit it from a devops perspective, and refactor deployment methodology/stack, etc. Maybe I'd benchmark it and realize its fine actually. Sometimes you just need to make your whole system work. We can obviously disagree with their priorities, their roadmap, the fact that the client isn't FOSS (I wish it was!), etc but no one can say that ollama doesn't work. It works. And like mchiang said above: its dead simple, on purpose. | ||||||||||||||
▲ | dcreater 7 days ago | parent [-] | |||||||||||||
But its effectively equally easy to do the same with llama.cpp, vllm or modular.. (any differences are small enough that they either shouldn't cause the human much work or can very easily be delegated to AI) | ||||||||||||||
|