▲ | ben_w 5 days ago | |||||||
Weirdly, you can not only do this, it somehow does actually catch some of its own mistakes. Not all of the mistakes, they generally still have a performance ceiling less than human experts (though even this disclaimer is still simplifying), but this kind of self-critique is basically what makes the early "reasoning" models one up over simple chat models: for the first-n :END: tokens, replace with "wait" and see it attempt other solutions and pick something usually better. | ||||||||
▲ | vrighter 4 days ago | parent [-] | |||||||
the "pick something usually better" sounds a lot like "and then draw the rest of the f*** owl" | ||||||||
|