| ▲ | NitpickLawyer 4 hours ago | ||||||||||||||||
It is an opinion piece. By a dude working as a "Professor of Pharmaceutical Technology and Biomaterials at the University of Ferrara". It has all the tropes of not understanding the underlying mechanisms, but repeating the common tropes. Quite ironic, considering what the author's intended "message" is. Jpeg -> jpeg -> jpeg bad. So llm -> llm -> llm must be bad, right? It reminds me of the media reception of that paper on model collapse. "Training on llm generated data leads to collapse". That was in 23 or 24? Yet we're not seeing any collapse, despite models being trained mainly on synthetic data for the past 2 years. That's not how any of it works. Yet everyone has an opinion on how bad it works. Jesus. It's insane how these kinds of opinion pieces get so upvoted here, while worth-while research, cool positive examples and so on linger in new with one or two upvotes. This has ceased to be a technical subject, and has moved to muh identity. | |||||||||||||||||
| ▲ | simonw 4 hours ago | parent | next [-] | ||||||||||||||||
Yeah, reading the other comments on this thread this is a classic example of that Hacker News (and online forums in general) thing where people jump on the chance to talk about a topic driven purely by the headline without engaging with the actual content. (I'm frequently guilty of that too.) | |||||||||||||||||
| |||||||||||||||||
| ▲ | PurpleRamen 4 hours ago | parent | prev [-] | ||||||||||||||||
> Yet we're not seeing any collapse, despite models being trained mainly on synthetic data for the past 2 years. Maybe because researchers learned from the paper to avoid the collapse? Just awareness alone often helps to sidestep a problem. | |||||||||||||||||
| |||||||||||||||||