▲ | pixl97 5 days ago | |||||||||||||||||||||||||
>find small perturbations that are undetectable to humans but produce a large change in model behavior. What artists don't realize by this they are just improving the models relative to human capabilities. The adversarial techniques like, for example making a stop sign look like something else, well likely be weeded out of the model by a convergence of model performance to average or above average human performance. | ||||||||||||||||||||||||||
▲ | pogue 5 days ago | parent [-] | |||||||||||||||||||||||||
How long until somebody comes up with another reCAPTCHA type system that forces users to click on images to identify them but that data is then used to verify training data for LLMs? (assuming this isn't happening already) | ||||||||||||||||||||||||||
|