| ▲ | SlinkyOnStairs 3 hours ago | |||||||||||||||||||||||||||||||||||||||||||||||||
> How else do you want companies to remove and prevent CSAM? Different situation. Facebook has to do CSAM moderation because it's a publishing platform. People will post CSAM on facebook, so they must do moderation. And "just don't have facebook" isn't a solution because every publication of any sort has to deal with this problem; Any newspaper accepting mail has this problem. (Albeit to a much more scaled down version) People were nailing obscene things to bulletin boards for all recorded history. --- In contrast, OpenAI has no such problem. It did not have CSAM pushed onto it, it actively collected such data itself. It could have, at any point before and after, simply stopped scraping all of the web indiscriminately and switched to using more curated sources of scraped data. The downside would be "worse LLMs" or "LLMs being created later", which is a perfectly acceptable compromise. --- This is not to say that genuine content flagging firms have no reason to curate such data & build tools to automatically flag content before human moderators have to. (But then they also shouldn't be outsourcing this and traumatizing contract workers for $2-3 an hour) But OpenAI is not such a firm. It's a general AI company. | ||||||||||||||||||||||||||||||||||||||||||||||||||
| ▲ | GrinningFool 2 hours ago | parent | next [-] | |||||||||||||||||||||||||||||||||||||||||||||||||
> traumatizing contract workers for $2-3 an hour) Is there an hourly rate at which this should be acceptable? | ||||||||||||||||||||||||||||||||||||||||||||||||||
| ||||||||||||||||||||||||||||||||||||||||||||||||||
| ▲ | deaux 2 hours ago | parent | prev | next [-] | |||||||||||||||||||||||||||||||||||||||||||||||||
> In contrast, OpenAI has no such problem. It did not have CSAM pushed onto it, it actively collected such data itself. It could have, at any point before and after, simply stopped scraping all of the web indiscriminately and switched to using more curated sources of scraped data. This is of course incredibly illegal, but megacorps (by valuation) and oligarchy members are above the law so who cares. I assume there could be a regulatory framework which can make this legal for an extremely specific purpose, but there is zero change that OpenAI was part of this/abiding by this in 2022, absolutely none. | ||||||||||||||||||||||||||||||||||||||||||||||||||
| ▲ | BobbyJo an hour ago | parent | prev | next [-] | |||||||||||||||||||||||||||||||||||||||||||||||||
> In contrast, OpenAI has no such problem. It did not have CSAM pushed onto it, it actively collected such data itself. It could have, at any point before and after, simply stopped scraping all of the web indiscriminately and switched to using more curated sources of scraped data. You've just thrown the garbage over your fence. Instead of OpenAI contracting Sama to classify CSAM, the "Curators" have to. At the end of the day, someone needs to classify it. If you say the platforms need to, and they miss some, and it ends up in OAI training data, OAI is going to be the entity paying the prices. | ||||||||||||||||||||||||||||||||||||||||||||||||||
| ▲ | fragmede 2 hours ago | parent | prev [-] | |||||||||||||||||||||||||||||||||||||||||||||||||
OpenAI runs ChatGPT where users submit text and photos and OpenAI generates and sends text and photos back. So users could be submitting CSAM. And yes, OpenAI could be generating CSAM. It's not limited to being a pull operation. What am I missing? | ||||||||||||||||||||||||||||||||||||||||||||||||||