| ▲ | Cadwhisker 5 hours ago | ||||||||||||||||
LMStudio? No, it's the easiest way to run am LLM locally that I've seen to the point where I've stopped looking at other alternatives. It's cross-platform (Win/Mac/Linux), detects the most appropriate GPU in your system and tells you whether the model you want to download will run within it's RAM footprint. It lets you set up a local server that you can access through API calls as if you were remotely connected to an online service. | |||||||||||||||||
| ▲ | vunderba 5 hours ago | parent [-] | ||||||||||||||||
FWIW, Ollama already does most of this: - Cross-platform - Sets up a local API server The tradeoff is a somewhat higher learning curve, since you need to manually browse the model library and choose the model/quantization that best fit your workflow and hardware. OTOH, it's also open-source unlike LMStudio which is proprietary. | |||||||||||||||||
| |||||||||||||||||