| ▲ | threethirtytwo 2 hours ago | |
What I read is the unlimited token count. You get the most out of this when having it run on an autonomous loop where your interaction is much more minimal? But pinging the thing every minute in a loop is going to terminate your token limit so running the LLM locally is the way to get infinite tokens. The problem is local models aren't as good as the ones in the cloud. I think the success stories are people who spent like 2-4k on a beefy system to run OpenClaw or these chatbots locally. The commands they run are, I assume like detailed versions of prompts that are essentially: "build my website." "Invest in stocks." And then watch it run for days. When using claude code it's essentially a partnership. You need to constantly manage it and curate it for safety but also so the token count doesn't go overboard. With a fully autonomous agent and unlimited token count you can assign it to tasks where this doesn't matter as much. Did the agent screw up and write bad code? The point is you can have the system prompt engage in self correction. | ||