| ▲ | toss1 2 days ago | |||||||||||||||||||||||||
The key point in the middle of the article. As AIs expand usage to larger numbers of lower-skilled coders whose lower ability to catch errors and provide feedback generates lower quality training data, the AIs are basically eating their own garbage, and the inevitable GIGO syndrome starts. >>But as inexperienced coders started turning up in greater numbers, it also started to poison the training data. >>AI coding assistants that found ways to get their code accepted by users kept doing more of that, even if “that” meant turning off safety checks and generating plausible but useless data. As long as a suggestion was taken on board, it was viewed as good, and downstream pain would be unlikely to be traced back to the source. | ||||||||||||||||||||||||||
| ▲ | Zababa 2 days ago | parent [-] | |||||||||||||||||||||||||
From what I understand model collapse/GIGO are not a problem in that labs generally know where the data comes from, so even if it causes problem in the long run you could filter it out. It's not like labs are forced to train models on the user outputs. | ||||||||||||||||||||||||||
| ||||||||||||||||||||||||||