| ▲ | pamelafox 7 hours ago | ||||||||||||||||||||||
This is why I only add information to AGENTS.md when the agent has failed at a task. Then, once I've added the information, I revert the desired changes, re-run the task, and see if the output has improved. That way, I can have more confidence that AGENTS.md has actually improved coding agent success, at least with the given model and agent harness. I do not do this for all repos, but I do it for the repos where I know that other developers will attempt very similar tasks, and I want them to be successful. | |||||||||||||||||||||||
| ▲ | viraptor 6 hours ago | parent | next [-] | ||||||||||||||||||||||
You can also save time/tokens if you see that every request starts looking for the same information. You can front-load it. | |||||||||||||||||||||||
| |||||||||||||||||||||||
| ▲ | averrous 7 hours ago | parent | prev | next [-] | ||||||||||||||||||||||
Agree. I also found out that rule discovery approach like this perform better. It is like teaching a student, they probably have already performed well on some task, if we feed in another extra rule that they already well verse at, it can hinder their creativity. | |||||||||||||||||||||||
| ▲ | imiric 6 hours ago | parent | prev [-] | ||||||||||||||||||||||
That's a sensible approach, but it still won't give you 100% confidence. These tools produce different output even when given the same context and prompt. You can't really be certain that the output difference is due to isolating any single variable. | |||||||||||||||||||||||
| |||||||||||||||||||||||