| ▲ | DeathArrow 5 hours ago | |
I see no mention of vLLM in the article. | ||
| ▲ | StrauXX 4 hours ago | parent [-] | |
vLLM isn't suitable for people running LLMs side-by-side with regular applications on their PC. It is very good at hosting LLMs for production on dedicated servers. For the prod usecase ollama/llamacpp are practically useless (but that's ok - it's not the projects goal to be). | ||