Remix.run Logo
segmondy 4 days ago

I know this is false, DeepSeekv3.1, GLM4.5, KimiK2-0905, Qwen-235B are all solid open models. Last night, I vibed rough 1300 lines of C server code in about an hour. 0 compilation error, ran without errors and got the job done. I want to meet this experienced programmer that can knock out 1300 lines of C code in an hour.

drusepth 4 days ago | parent | next [-]

Are 235B models classified as local LLMs? I guess they probably are, but others in this thread are probably looking more toward 20B-30B models and sizes that generally fit on the RAM you'd expect in average or slightly-higher-end hardware.

My beefy 3D gamedev workstation with a 4090 and 128GB RAM can't even run a 235B model unless it's extremely quantized (and even then, only at like single-digit tokens/minute).

codazoda 4 days ago | parent | prev | next [-]

How much machine do you have to be able to run Qwen-235B locally?

oblio 4 days ago | parent | prev | next [-]

Can you run any of those models without $20 000 worth of hardware that uses as much power and makes as much noise as a small factory?

segmondy 2 days ago | parent [-]

I run them with under $3,000 hardware and inference is about 500-600watts with no noise.

nomel 4 days ago | parent | prev | next [-]

Without knowing what you were doing with that 1300 lines of code, there's not much insight that can be had from this.

brookst 3 days ago | parent | prev [-]

I’m a mediocre C programmer on my best day and I assure you a highly competent programmer could probably use 200 lines of code to do what I achieve in 1300.

Just counting lines is not a good proxy for how much effort it would take a good programmer.

(And I am 100% pro LLM coding, just saying this isn’t a great argument)