Remix.run Logo
verdverm 6 hours ago

We are exiting a hype cycle, well into the adoption curve. Subscriptions were never going to last.

My next step is going to be evaluating open and local models to see if they are sufficiently close to par with frontier models.

My hope is that the end of seat based pricing comes with this tech cycle. I was looking for document signing provider that doesn't charge a monthly, I only need a few docs a year.

alifeinbinary 6 hours ago | parent | next [-]

I'm developing software in this area right now, so I try a lot of the new models. They're not even close for coding tasks. It basically comes down to 26b parameters vs 1T parameters / quantisation / smaller context sizs, there's no comparison. However, for agentic work, tool calling, text summarisation, local LLMs can be quite capable. Workloads that run as background tasks where you're not concerned about TTFB, cold starts, tok/s etc., this is where local AI is useful.

If you have an M processor then I would recommend that you ditch Ollama because it performs slowly. We get double or triple tok/s using omlx or vmlx, respectively, but vmlx doesn't have extensive support for some models like gpt-oss.

AstroBen 5 hours ago | parent | next [-]

Kimi K2.5 (as an example) is an open model with 1T params. I don't see a reason it has to be local for most use cases- the fact that it's open is what's important.

verdverm 4 hours ago | parent | prev [-]

first session with gemma4:31b looks pretty good, like it may actually be up to coding tasks like gemini-3-flash levels

you can tell gemma4 comes from gemini-3

__mharrison__ 6 hours ago | parent | prev [-]

I recently experimented creating a Python library from scratch with Codex. After I was done, I took the PRD and Task list that was generated and fed them to opencode with Qwen 3.5 running locally.

Opencode was able to create the library as well. It just took about 2x longer.

selectodude 6 hours ago | parent [-]

Which version of Qwen 3.5 did you use?

verdverm 6 hours ago | parent [-]

which quant as well

__mharrison__ 3 hours ago | parent [-]

Not at my computer now, either 27 or 35b not quantized.

Next week I will be trying qwopus 27b.