Remix.run Logo
locknitpicker 2 days ago

> It's pointless to write a whole article about how model collapse is actually happening and isn't just a theoretical concern with no evidence that model collapse is actually happening.

Except perhaps the link to article on the peer-reviewed paper that describes the problem in detail.

https://www.cs.ox.ac.uk/news/2356-full.html

> Researchers at Oxford and Cambridge published work on this back in 2023, showing how iterative training on synthetic data leads to progressive degradation.

Legend2440 2 days ago | parent [-]

This is a toy example of how it could happen, in an artificial setting where you train entirely on generated outputs many times in a row.

It does not say that it is happening in production LLMs. It is a theoretical concern right now.