▲ | muzani 4 days ago | |||||||
I don't have a source for this (there's probably no sources from anything back then) but anecdotally, someone at an AI/ML talk said they just added more data and quality went up. Doubling the data doubled the quality. With other breakthroughs, people saw diminishing gains. It's sort of why Sam back then tweeted that he expected the amount of intelligence to double every N years. I have the feeling they kept on this until GPT-4o (which was a different kind of data). | ||||||||
▲ | robrenaud 4 days ago | parent [-] | |||||||
The input size to output quality mapping is not linear. This is why we are in the regime of "build nuclear power plants to power datacenters". Fixed size improvements in loss require exponential increases in parameters/compute/data. | ||||||||
|