| ▲ | Hasslequest 2 days ago | |
I run Qwen-3.5 based LLMs in the 20-35 parameter range on my laptop's iGPU and it works great for my use case, which includes coding, search, reasoning, and general tasks. Gemma3 is good too. There are ones that are distilled with better reasoning models or abliterated for whatever you need, and the multimodal features work... fine. Just started running local LLMs this week, and it is pretty much overkill for what anyone in my family needs. All it really lacks is some tools for it to use, which I am putting together now. To be fair, the best model I have used is claude sonnet. I don't really know what I am missing with opus. | ||