| ▲ | zozbot234 7 hours ago | |||||||
> We will run out of additional material to train on This sounds a bit silly. More training will generally result in better modeling, even for a fixed amount of genuine original data. At current model sizes, it's essentially impossible to overfit to the training data so there's no reason why we should just "stop". | ||||||||
| ▲ | _0ffh 5 hours ago | parent | next [-] | |||||||
You'd be surprised how quickly improvement of autoregressive language models levels off with epoch count (though, admittedly, one epoch is a LOT). Diffusion language models otoh indeed keep profiting for much longer, fwiw. | ||||||||
| ▲ | pvab3 6 hours ago | parent | prev [-] | |||||||
I'm just talking about text generated by human beings. You can keep retraining with more parameters on the same corpus | ||||||||
| ||||||||