| ▲ | TheBlight 11 hours ago | |||||||
"Scaling" is going to eventually apply to the ability to run more and higher fidelity simulations such that AI can run experiments and gather data about the world as fast and as accurately as possible. Pre-training is mostly dead. The corresponding compute spend will be orders of magnitude higher. | ||||||||
| ▲ | alyxya 11 hours ago | parent [-] | |||||||
That's true, I expect more inference time scaling and hybrid inference/training time scaling when there's continual learning rather than scaling model size or pretraining compute. | ||||||||
| ||||||||