|
| ▲ | duffyjp 6 hours ago | parent | next [-] |
| Nothing. This summer I set up a dual 16GB GPU / 64GB RAM system and nothing I could run was even remotely close. Big models that didn't fit on 32gb VRAM had marginally better results but were at least of magnitude slower than what you'd pay for and still much worse in quality. I gave one of the GPUs to my kid to play games on. |
| |
| ▲ | Tostino 5 hours ago | parent [-] | | Yup, even with 2x 24gb GPUs, it's impossible to get anywhere close to the big models in terms of quality and speed, for a fraction of the cost. | | |
| ▲ | mirekrusin 2 hours ago | parent [-] | | I'm running unsloth/GLM-4.7-Flash-GGUF:UD-Q8_K_XL via llama.cpp on 2x 24G 4090s which fits perfectly with 198k context at 120 tokens/s – the model itself is really good. |
|
|
|
| ▲ | medvezhenok 7 hours ago | parent | prev | next [-] |
| Short answer: there is none. You can't get frontier-level performance from any open source model, much less one that would work on an M3 Pro. If you had more like 200GB ram you might be able to run something like MiniMax M2.1 to get last-gen performance at something resembling usable speed - but it's still a far cry from codex on high. |
|
| ▲ | mittermayr 7 hours ago | parent | prev | next [-] |
| at the moment, I think the best you can do is qwen3-coder:30b -- it works, and it's nice to get some fully-local llm coding up and running, but you'll quickly realize that you've long tasted the sweet forbidden nectar that is hosted llms. unfortunately. |
|
| ▲ | evilduck 5 hours ago | parent | prev | next [-] |
| They are spending hundreds of billions of dollars on data centers filled with GPUs that cost more than an average car and then months on training models to serve your current $20/mo plan. Do you legitimately think there's a cheaper or free alternative that is of the same quality? I guess you could technically run the huge leading open weight models using large disks as RAM and have close to the "same quality" but with "heat death of the universe" speeds. |
|
| ▲ | tosh 2 hours ago | parent | prev | next [-] |
| 18gb RAM it is a bit tight with 32gb RAM: qwen3-coder and glm 4.7 flash are both impressive 30b parameter models not on the level of gpt 5.2 codex but small enough to run locally (w/ 32gb RAM 4bit quantized) and quite capable but it is just a matter of time I think until we get quite capable coding models that will be able to run with less RAM |
|
| ▲ | Mashimo 7 hours ago | parent | prev | next [-] |
| A local model with 18GB of ram that has the same quality has codex high? Yeah, nah mate. The best could be GLN 4.7 Flash, and I doubt it's close to what you want. |
|
| ▲ | atwrk 7 hours ago | parent | prev | next [-] |
| "run" as in run locally? There's not much you can do with that little RAM. If remote models are ok you could have a look at MiniMax M2.1 (minimax.io) or GLM from z.ai or Qwen3 Coder. You should be able to use all of these with your local openai app. |
|
| ▲ | jgoodhcg 6 hours ago | parent | prev | next [-] |
| Z.ai has glm-4.7. Its almost as good for about $8/mo. |
| |
| ▲ | margorczynski 6 hours ago | parent [-] | | Not sure if it's me but at least for my use cases (software devl, small-medium projects) Claude Opus + Claude Code beats by quite a margin OpenCode + GLM 4.7. At least for me Claude "gets it" eventually while GLM will get stuck in a loop not understanding what the problem is or what I expect. | | |
| ▲ | zamalek 5 hours ago | parent [-] | | Right, GLM is close But not close enough. If I have to spend $200 for Opus fallback i may as well not use it always. Still an unbelievable option if $200 is a luxury, the price-per-quality is absurd. |
|
|
|
| ▲ | marcd35 5 hours ago | parent | prev [-] |
| antigravity is solid and has a generous free tier. |