| ▲ | chongli 4 hours ago | |||||||||||||
New models literally do get worse after launch, due to optimization. If you charted performance over time, it'd look like a sawtooth, with a regular performance drop during each optimization period. That's the dirty secret with all of this stuff: "state of the art" models are unprofitable due to high cost of inference before optimization. After optimization they still perform okay, but way below SOTA. It's like a knife that's been sharpened until razor sharp, then dulled shortly after. | ||||||||||||||
| ▲ | girvo 3 hours ago | parent | next [-] | |||||||||||||
> If you charted performance over time, it'd look like a sawtooth People have, though, and it doesn't show that. I think it's more people getting hit by the placebo effect, the novelty effect, followed by the models by-definition non-determinism leading people to say things like "the model got worse". | ||||||||||||||
| ▲ | gobdovan 3 hours ago | parent | prev [-] | |||||||||||||
Is this insider info? The 'charted performance' caught my eye instantly. Couple things I find odd tho: why sawtooth? it would likely be square waves, as I'd imagine they roll down the cost-saving version quite fast per cohort. Also, aren't they unprofitable either way? Why would they do it for 'profitability'? | ||||||||||||||
| ||||||||||||||