| ▲ | moqizhengz 4 hours ago |
| Running 3.5 9B on my ASUS 5070ti 16G with lm studio gives a stable ~100 tok/s.
This outperforms the majority of online llm services and the actual quality of output matches the benchmark.
This model is really something, first time ever having usable model on consumer-grade hardware. |
|
| ▲ | smokel 21 minutes ago | parent | next [-] |
| > This outperforms the majority of online llm services I assume you mean outperforms in speed on the same model, not in usability compared to other more capable models. (For those who are getting their hopes up on using local LLMs to be any replacement for Sonnet or Opus.) |
|
| ▲ | yangikan 3 hours ago | parent | prev | next [-] |
| Do you point claude code to this? The orchestration seems to be very important. |
| |
| ▲ | badgersnake 22 minutes ago | parent | next [-] | | I’ve tried it on Claude code,
Found it to be fairly crap. It got stuck in a loop doing the wrong thing and would not be talked out of it. I’ve found this bug that would stop it compiling right after compiling it, that sort of thing. Also seemed to ignore fairly simple instructions in CLAUDE.md about building and running tests. | |
| ▲ | teaearlgraycold 2 hours ago | parent | prev [-] | | I loaded Qwen into LM Studio and then ran Oh My Pi. It automatically picked up the LM Studio API server. For some reason the 35B A3B model had issues with Oh My Pi's ability to pass a thinking parameter which caused it to crash. 27B did not have that issue for me but it's much slower. Here's how I got the 35B model to work: https://gist.github.com/danthedaniel/c1542c65469fb1caafabe13... The 35B model is still pretty slow on my machine but it's cool to see it working. |
|
|
| ▲ | throwdbaaway 3 hours ago | parent | prev | next [-] |
| There are Qwen3.5 27B quants in the range of 4 bits per weight, which fits into 16G of VRAM. The quality is comparable to Sonnet 4.0 from summer 2025. Inference speed is very good with ik_llama.cpp, and still decent with mainline llama.cpp. |
| |
| ▲ | codemog 2 hours ago | parent | next [-] | | Can someone explain how a 27B model (quantized no less) ever be comparable to a model like Sonnet 4.0 which is likely in the mid to high hundreds of billions of parameters? Is it really just more training data? I doubt it’s architecture improvements, or at the very least, I imagine any architecture improvements are marginal. | | |
| ▲ | revolvingthrow 13 minutes ago | parent | next [-] | | It doesn’t. I’m not sure it outperforms chatgpt 3 | |
| ▲ | spwa4 11 minutes ago | parent | prev | next [-] | | The short answer is that there are more things that matter than parameter count, and we are probably nowhere near the most efficient way to make these models. | |
| ▲ | otabdeveloper4 an hour ago | parent | prev [-] | | There's diminishing returns bigly when you increase parameter count. The sweet spot isn't in the "hundreds of billions" range, it's much lower than that. Anyways your perception of a model's "quality" is determined by careful post-training. | | |
| ▲ | codemog an hour ago | parent | next [-] | | Interesting. I see papers where researchers will finetune models in the 7 to 12b range and even beat or be competitive with frontier models. I wish I knew how this was possible, or had more intuition on such things. If anyone has paper recommendations, I’d appreciate it. | |
| ▲ | zozbot234 an hour ago | parent | prev [-] | | More parameters improves general knowledge a lot, but you have to quantize higher in order to fit in a given amount of memory, which if taken to extremes leads to erratic behavior. For casual chat use even Q2 models can be compelling, agentic use requires more regularization thus less quantized parameters and lowering the total amount to compensate. |
|
| |
| ▲ | zozbot234 2 hours ago | parent | prev | next [-] | | With MoE models, if the complete weights for inactive experts almost fit in RAM you can set up mmap use and they will be streamed from disk when needed. There's obviously a slowdown but it is quite gradual, and even less relevant if you use fast storage. | |
| ▲ | teaearlgraycold 2 hours ago | parent | prev [-] | | Qwen3.5 35B A3B is much much faster and fits if you get a 3 bit version. How fast are you getting 27B to run? On my M3 Air w/ 24GB of memory 27B is 2 tok/s but 35B A3B is 14-22 tok/s which is actually usable. | | |
| ▲ | throwdbaaway 35 minutes ago | parent | next [-] | | Using ik_llama.cpp to run a 27B 4bpw quant on a RTX 3090, I get 1312 tok/s PP and 40.7 tok/s TG at zero context, dropping to 1009 tok/s PP and 36.2 tok/s TG at 40960 context. 35B A3B is faster but didn't do too well in my limited testing. | |
| ▲ | ece 2 hours ago | parent | prev [-] | | The 27B is rated slightly higher for SWE-bench. |
|
|
|
| ▲ | lukan 3 hours ago | parent | prev [-] |
| What exact model are you using? I have a 16GB GPU as well, but have never run a local model so far. According to the table in the article, 9B and 8-bit -> 13 GB and 27B and 3-bit seem to fit inside the memory. Or is there more space required for context etc? |
| |
| ▲ | vasquez an hour ago | parent [-] | | It depends on the task, but you generally want some context. These models can do things like OCR and summarize a pdf for you, which takes a bit of working memory. Even more so for coding CLIs like opencode-ai, qwen code and mistral ai. Inference engines like llama.cpp will offload model and context to system ram for you, at the cost of performance. A MoE like 35B-A3B might serve you better than the ones mentioned, even if it doesn't fit entirely on the GPU. I suggest testing all three. Perhaps even 122-A10B if you have plenty of system ram. Q4 is a common baseline for simple tasks on local models. I like to step up to Q5/Q6 for anything involving tool use on the smallish models I can run (9B and 35B-A3B). Larger models tolerate lower quants better than small ones, 27B might be usable at 3 bpw where 9B or 4B wouldn't. You can also quantize the context. On llama.cpp you'd set the flags -fa on, -ctk x and ctv y. -h to see valid parameters. K is more sensitive to quantization than V, don't bother lowering it past q8_0. KV quantization is allegedly broken for Qwen 3.5 right now, but I can't tell. |
|