| ▲ | locknitpicker 2 days ago | |
> It's pointless to write a whole article about how model collapse is actually happening and isn't just a theoretical concern with no evidence that model collapse is actually happening. Except perhaps the link to article on the peer-reviewed paper that describes the problem in detail. https://www.cs.ox.ac.uk/news/2356-full.html > Researchers at Oxford and Cambridge published work on this back in 2023, showing how iterative training on synthetic data leads to progressive degradation. | ||
| ▲ | Legend2440 2 days ago | parent [-] | |
This is a toy example of how it could happen, in an artificial setting where you train entirely on generated outputs many times in a row. It does not say that it is happening in production LLMs. It is a theoretical concern right now. | ||