| ▲ | jlokier 4 days ago | |||||||
I use a Macbook Pro with 128GB RAM "unified memory" that's available to both CPU and GPU. It's slower than a rented Nvidia GPU, but usable for all the models I've tried (even gpt-oss-120b), and works well in a coffee shop on battery and with no internet connection. I use Ollama to run the models, so can't run the latest until they are ported to the Ollama library. But I don't have much time for tinkering anyway, so I don't mind the publishing delay. | ||||||||
| ▲ | anon373839 4 days ago | parent | next [-] | |||||||
I’d strongly advise ditching Ollama for LM Studio, and using MLX versions of the models. They run quite a bit faster on Apple Silicon. Also, LM Studio is much more polished and feature rich than Ollama. | ||||||||
| ||||||||
| ▲ | MaxMatti 4 days ago | parent | prev [-] | |||||||
How's the battery holding up during vibe coding sessions or occasional LLM usage? I've been thinking about getting a MacBook or a laptop with a similar Ryzen chip specifically for that reason. | ||||||||