| ▲ | Ask HN: How do you catch regressions when you change your AI agent's prompt? | |
| 1 points by yakshithk_ 10 hours ago | 1 comments | ||
i've been building agents for a while and kept running into the same problem: change the sys prompt, swap models, or tweak for the agent calls a tool — and something subtly breaks. The only way I found out was manually running it or a user reporting it. | ||
| ▲ | SashaMApps 5 hours ago | parent [-] | |
Any AI is is non-deterministic by nature, this is more about statistics, but not dedicated test cases. So, even if you swich to the model which is better in metric results you still cannot guarantee it will be better in your specific use case or do the same. What you can probably do is to automate your important test scenarious and run each every time you change model or system prompt | ||