| ▲ | zahlman 7 hours ago | |||||||
> Otherwise, you'd expect the effect to be similar to telling the model 'make a plan before acting, make not mistakes'. Have there not been previous iterations of these tools where such techniques were actually effective? | ||||||||
| ▲ | gwern 6 hours ago | parent | next [-] | |||||||
But that's a reason you should expect it to stop working soon, just like all the older tricks like "my grandmother will die". If you have a universal 'blind' prompt which can increase performance a little bit... the AI labs can just toss that into the training loop to teach the model to do it automatically, whatever 'it' was, like 'trying harder' or 'writing down a useful idea'. And then the prompt stops working because the next generations do it by default. (This also suggests that you should expect them to generally be bad at judging novel self-generated prompts/skills - if they could judge those, they would already be using them! There is a generator-verifier gap, but it is already exploited heavily during post-training and not much low-hanging fruit left there.) | ||||||||
| ||||||||
| ▲ | rahimnathwani 7 hours ago | parent | prev [-] | |||||||
Yes, but this paper studied recent models. | ||||||||