Remix.run Logo
DeathArrow 5 hours ago

I see no mention of vLLM in the article.

StrauXX 4 hours ago | parent [-]

vLLM isn't suitable for people running LLMs side-by-side with regular applications on their PC. It is very good at hosting LLMs for production on dedicated servers. For the prod usecase ollama/llamacpp are practically useless (but that's ok - it's not the projects goal to be).