▲ | ACCount37 3 days ago | ||||||||||||||||
The raw model scale is not increasing by much lately. AI companies are constrained by what fits in this generation of hardware, and waiting for the next generation to become available. Models that are much larger than the current frontier are still too expensive to train, and far too expensive to serve them en masse. In the meanwhile, "better data", "better training methods" and "more training compute" are the main ways you can squeeze out more performance juice without increasing the scale. And there are obvious gains to be had there. | |||||||||||||||||
▲ | robwwilliams 3 days ago | parent | next [-] | ||||||||||||||||
The jump to 1 million token length context for Sonnet 4 plus access to internet has been a game-changer for me. And somebody should remind Anthropic leadership to at least mirror Wikipedia; better yet support Wikipedia actively. All of the big AI players have profited from Wikipedia, but have they given anything back, or are they just parasites on FOSS and free data? | |||||||||||||||||
▲ | xnx 3 days ago | parent | prev [-] | ||||||||||||||||
> AI companies are constrained by what fits in this generation of hardware, and waiting for the next generation to become available. Does this apply to Google that is using custom built TPUs while everyone else uses stock Nvidia? | |||||||||||||||||
|