| ▲ | zozbot234 4 hours ago | |||||||||||||
Mind you, a 30B model (3B active) is not going to be comparable to Opus. There are open models that are near-SOTA but they are ~750B-1T total params. That's going to require substantial infrastructure if you want to use them agentically, scaled up even further if you expect quick real-time response for at least some fraction of that work. (Your only hope of getting reasonable utilization out of local hardware in single-user or few-users scenarios is to always have something useful cranking in the background during downtime.) | ||||||||||||||
| ▲ | pitched 4 hours ago | parent | next [-] | |||||||||||||
For a business with ten or more engineers/people-using-ai, it might still make sense to set this up. For an individual though, I can’t imagine you’d make it through to positive ROI before the hardware ages out. | ||||||||||||||
| ||||||||||||||
| ▲ | wuschel 3 hours ago | parent | prev | next [-] | |||||||||||||
What near SOTA open models are you referring to? | ||||||||||||||
| ▲ | cyberax 3 hours ago | parent | prev [-] | |||||||||||||
I'm backing up a big dataset onto tapes, so I wanted to automate it. I have an idle 64Gb VRAM setup in my basement, so I decided to experiment and tasked it with writing an LTFS implementation. LTFS is an open standard for filesystems for tapes, and there's an implementation in C that can be used as the baseline. So far, Qwen 3.6 created a functionally equivalent Golang implementation that works against the flat file backend within the last 2 days. I'm extremely impressed. | ||||||||||||||
| ||||||||||||||