| ▲ | SOLAR_FIELDS 4 hours ago | |||||||||||||||||||||||||
I think the key you point out is something that is worth observing more generically - if the LLM hits a wall it’s first inkling is not to step back and understand why the wall exists and then change course, its first inkling is to continue assisting the user on its task by any means possible and so it’s going to instead try to defeat it in any way possible. I see the is all the time when it hits code coverage constraints, it would much rather just lower thresholds than actually add more coverage. I experimented with hooks a lot over the summer, these kind of deterministic hooks that run before commit, after tool call, after edit, etc and I found they are much more effective if you are (unsurprisingly) able to craft and deliver a concise, helpful error message to the agent on the hook failure feedback. Even just giving it a good howToFix string in the error return isn’t enough, if you flood the response with too many of those at once the agent will view the task as insurmountable and start seeking workarounds instead. | ||||||||||||||||||||||||||
| ▲ | AdieuToLogic 3 hours ago | parent [-] | |||||||||||||||||||||||||
> ... if the LLM hits a wall it’s first inkling is not to step back and understand why the wall exists and then change course, its first inkling is ... LLM's do not "understand why." They do not have an "inkling." Claiming they do is anthropomorphizing a statistical token (text) document generator algorithm. | ||||||||||||||||||||||||||
| ||||||||||||||||||||||||||