| ▲ | llagerlof 4 hours ago | ||||||||||||||||||||||||||||||||||
What’s new regarding llmfit? | |||||||||||||||||||||||||||||||||||
| ▲ | karmakaze an hour ago | parent | next [-] | ||||||||||||||||||||||||||||||||||
This has a web version[0] which I wish they'd host on a free site. [0] https://github.com/AlexsJones/llmfit/tree/main/llmfit-web Edit: I tried to deploy a snapshot of the llmfit-web files on Netlify but it seems to need/want to talk to a backend[1] | |||||||||||||||||||||||||||||||||||
| ▲ | rvz 4 hours ago | parent | prev | next [-] | ||||||||||||||||||||||||||||||||||
Other than it (whichllm) being written in Python, nothing else. I just use llmfit. | |||||||||||||||||||||||||||||||||||
| ▲ | andyyyy64 2 hours ago | parent | prev [-] | ||||||||||||||||||||||||||||||||||
Fair question. llmfit answers "will this model fit in my memory?" — it's a fit/size calculator, and a good one. whichllm answers a different question: "of the models that fit, which is actually best?" It pulls candidates, then ranks them by merged real benchmarks (LiveBench / Artificial Analysis / Aider / Arena ELO / Open LLM Leaderboard) with a recency penalty, so a newer 27B beats an older 32B even though both fit — on a 24GB card it puts Qwen3.6-27B above Qwen3-32B on benchmarks, not size. If "biggest that fits" is the answer you want, llmfit is the simpler tool and Python won't matter to you. If you want "which fitting model is worth running," that ranking layer is the whole reason whichllm exists. Different jobs — I'd genuinely send fit-only users to llmfit. | |||||||||||||||||||||||||||||||||||
| |||||||||||||||||||||||||||||||||||