▲ | prisenco 6 days ago | |||||||||||||||||||||||||
Yes, exactly. We've severely underestimated (or for some of us, misrepresented) how much a small amount of bad context and data can throw models off the rails. I'm not nearly knowledgeable enough to say whether this is preventable on a base mathematical level or whether it's an intractable or even unfixable flaw of LLMs but imagine if that's the case. | ||||||||||||||||||||||||||
▲ | JoshTriplett 6 days ago | parent | next [-] | |||||||||||||||||||||||||
Closely related concept: https://en.wikipedia.org/wiki/Waluigi_effect | ||||||||||||||||||||||||||
| ||||||||||||||||||||||||||
▲ | derbOac 6 days ago | parent | prev [-] | |||||||||||||||||||||||||
My sense is this is reflective of a broader problem with overfitting or sensitivity (my sense is they are flip sides of the same coin). Ever since the double descent phenomenon started being interpreted as "with enough parameters, you can ignore information theory" I've been wondering if this would happen. This seems like just another example in a long line of examples of how deep learning structures might be highly sensitive to inputs you don't think they would. | ||||||||||||||||||||||||||
|