| ▲ | Probably quit your job if you're asking Reddit(datastream.substack.com) |
| 5 points by racketracer 7 hours ago | 5 comments |
| |
|
| ▲ | data4lyfe 6 hours ago | parent | next [-] |
| Most people likely cannot quit a high paying job when their identity is also wrapped in how much they’re earning. I see this a lot from all of the newly minted AI millionaires. |
|
| ▲ | panny 7 hours ago | parent | prev [-] |
| In other news, 40% of your LLM's training data is reddit posts. Derive from that what you will. |
| |
| ▲ | perrygeo 6 hours ago | parent [-] | | Where did you get 40%. I'm sure reddit content is all in the training set but that seems high. If it is that high, reddit comments seems like a ripe target for LLM poisoning. | | |
| ▲ | racketracer 6 hours ago | parent [-] | | What is LLM poisoning? You're saying if I create a prompt that says "Classify this comment if it's XYZ or asking for ABC" that the LLM will just not do it correctly because it's trained on Reddit? | | |
| ▲ | perrygeo 5 hours ago | parent [-] | | LLM poisoning refers to feeding the model false information during training. Anti-AI folks are openly talking about intentionally flooding the internet with garbage to reduce the quality of the models. Reddit just provides a convenient and barely moderated forum for them to spread misinformation. And it doesn't take much: https://www.anthropic.com/research/small-samples-poison |
|
|
|