Remix.run Logo
zkmon 8 hours ago

Yesterday was a realization point for me. I gave a simple extraction task to Claude code with a local LLM and it "whirred" and "purred" for 10 minutes. Then I submitted the same data and prompt directly to model via llama_cpp chat UI and the model single-shotted it in under a minute. So obviously something wrong with coding agent or the way it is talking to LLM.

Now I'm looking for an extremely simple open-source coding agent. Nanocoder doesn't seem install on my Mac and it brings node-modules bloat, so no. Opencode seems not quite open-source. For now, I'm doing the work of coding agent and using llama_cpp web UI. Chugging it along fine.

syhol 8 hours ago | parent | next [-]

https://pi.dev/ seems popular, whats not open source about opencode? The repo has an MIT License.

xlii 2 hours ago | parent | next [-]

+1 for pi. I used claude and opencode but pi is the first agent tool that made me excited about the whole thing.

tfrancisl 8 hours ago | parent | prev | next [-]

Some people believe only copyleft licenses are open source. They're right on principle, wrong in (legal) practice.

steveklabnik 7 hours ago | parent [-]

They're not even right on principle: https://www.gnu.org/licenses/license-list.html

Even the FSF recognizes that non-copyleft licenses still follow the Freedoms, and therefore are still Free Software.

fortyseven 7 hours ago | parent | prev | next [-]

Been LOVING Pi so far!

zkmon 7 hours ago | parent | prev [-]

Maybe it's just my feeling. It asks to update/upgrade continuously.

BeetleB 5 hours ago | parent [-]

It's completely open source, but is under heavy continual development (likely a lot of AI coding).

On launch, it checks for updates and autoupdates.

SyneRyder 7 hours ago | parent | prev | next [-]

Probably a silly idea, but I'll throw it into the mix - have your current AI build one for you. You can have exactly the coding agent you want, especially if you're looking for "extremely simple".

I got annoyed enough with Anthropic's weird behavior this week to actually try this, and got something workable up & running in a few days. My case was unique: there's no Claude Code for BeOS, or my older / ancient Macs, so it was easier to bootstrap & stitch something together if I really wanted an agentic coding agent on those platforms. You'll learn a lot about how models actually work in the process too, and how much crazy ridiculous bandaid patching is happening Claude Code. Though you might also appreciate some of the difficulties that the agent / harnesses have to solve too. (And to be clear, I'm still using CC when I'm on a platform that supports it.)

As for the llama_cpp vs Claude Code delays - I've run into that too. My theory is API is prioritized over Claude Code subscription traffic. API certainly feels way faster. But you're also paying significantly more.

appcustodian2 8 hours ago | parent | prev | next [-]

Just in case it didn't occur to you already, you can just build whatever coding agent you want. They're pretty simple

btbuildem 5 hours ago | parent | prev | next [-]

You'd figure by now we would have something between a TUI and an IDE.

btbuildem 7 hours ago | parent | prev | next [-]

You can run CC with local models, it's pretty straightforward. I've done this with vLLM + a thin shim to change the endpoint syntax.

jedisct1 8 hours ago | parent | prev | next [-]

Swival is not bloated and was specifically made for local agents: https://swival.dev

pferdone 8 hours ago | parent [-]

pi.dev as well

banditelol 8 hours ago | parent | prev | next [-]

what model you used with llama_cpp?

zkmon 7 hours ago | parent [-]

Qwen3.6-35B quant-4 gguf

enraged_camel 8 hours ago | parent | prev [-]

I use both Cursor and Claude Code, and yes, the latter is noticeably slower with the same model at the same settings.

However, it's hard to justify Cursor's cost. My bill was $1,500/mo at one point, which is what encouraged me to give CC a try.