▲ | sottol 6 days ago | ||||||||||||||||||||||
Maybe that's true for absolute arm-chair-engineering outsiders (like me) but these models are in training for months, training data is probably being prepared year(s) in advance. These models have a knowledge cut-off in 2024 - so they have been in training for a while. There's no way sama did not have a good idea that this non-COT was in the pipeline 2 months ago. It was probably finished training then and undergoing evals. Maybe 1. he's just doing his job and hyping OpenAI's competitive advantages (afair most of the competition didn't have decent COT models in Feb), or 2. something changed and they're releasing models now that they didn't intend to release 2 months ago (maybe because a model they did intend to release is not ready and won't be for a while), or 3. COT is not really as advantageous as it was deemed to be 2+ months ago and/or computationally too expensive. | |||||||||||||||||||||||
▲ | fragmede 6 days ago | parent [-] | ||||||||||||||||||||||
With new hardware from Nvidia announced coming out, those months turn into weeks. | |||||||||||||||||||||||
|