| ▲ | noosphr an hour ago | |
I've been at this longer than most. After three major generations of models the "intuition" I've build isn't about what AI can do, but about what a specific model family can do. No one cares what the gotchas in gpt3 are because it's a stupid model. In two years no one will care what they were for gpt5 or Claude 4 for the same reason. We currently have the option of wasting months of our lives to get good at a specific model, or burn millions to try and get those models to do things by themselves. Neither option is viable long term. | ||
| ▲ | CuriouslyC an hour ago | parent [-] | |
My philosophy is to try and model the trajectories of these systems and build rigging around where the curve is flat (e.g. models have been producing big balls of mud since the beginning and this hasn't improved meaningfully). Models also have a strong mean bias that I don't expect to go away any time soon. Trying to outsmart the models at core behaviors over time is asking to re-learn the bitter lesson though. | ||