Remix.run Logo
QuadrupleA 5 hours ago

Claude Code's primarily optimized for burning as many tokens as possible.

redman25 41 minutes ago | parent | next [-]

It’s mainly the benchmarks that have encouraged that. The more tokens they crank out the more likely the answer is to be somewhere in the output.

tartoran 5 hours ago | parent | prev | next [-]

Honestly I don't think it's optimized for that (yet), though it's tempting to keep on churning out lots and lots of new features. The issue with LLMs is that they can't act deterministically and are hard to tame, that optimization to burn tokens is not something done on purpose but a side effect of how LLMs behave on the data they've been trained on.

ysleepy 4 hours ago | parent [-]

set the temperature=0 and it is (pretty much) deterministic.

But I assume you mean predictable in the sense of reacting simiarly to similar inputs.

tartoran 25 minutes ago | parent [-]

I havent played with temperature much. Did you? Do you ever make use of temp=0?

arcanemachiner 5 hours ago | parent | prev [-]

That's OpenCode. The model is Claude Opus, which is probably RL'ed pretty heavily to work with Claude Code. So it's a little less surprising to see it bungle the intentions since it's running in another harness. Still laughable though.

RL - reinforcement learning