| ▲ | sheept 3 hours ago | |||||||||||||||||||||||||||||||
This feels completely speculative: there's no measure of whether this approach is actually effective. Personally, I'm skeptical: - Having the agent look up the JSON schemas and skills to use the CLI still dumps a lot of tokens into its context. - Designing for AI agents over humans doesn't seem very future proof. Much of the world is still designed for humans, so the developers of agents are incentivized to make agents increasingly tolerate human design. - This design is novel and may be fairly unfamiliar in the LLM's training data, so I'd imagine the agent would spend more tokens figuring this CLI out compared to a more traditional, human-centered CLI. | ||||||||||||||||||||||||||||||||
| ▲ | gck1 3 hours ago | parent | next [-] | |||||||||||||||||||||||||||||||
Yeah, people seem to forget one of the L's in LLM stands for Language, and human language is likely the largest chunk in training data. A cli that is well designed for humans is well designed for agents too. The only difference is that you shouldn't dump pages of content that can pollute context needlessly. But then again, you probably shouldn't be dumping pages of content for humans either. | ||||||||||||||||||||||||||||||||
| ||||||||||||||||||||||||||||||||
| ▲ | an hour ago | parent | prev | next [-] | |||||||||||||||||||||||||||||||
| [deleted] | ||||||||||||||||||||||||||||||||
| ▲ | magospietato 3 hours ago | parent | prev [-] | |||||||||||||||||||||||||||||||
Surely the skill for a cli tool is a couple of lines describing common usage, and a description of the help system? | ||||||||||||||||||||||||||||||||
| ||||||||||||||||||||||||||||||||