| ▲ | caditinpiscinam 3 hours ago | |
I think you raise a valid point about the bias inherent in these models. I'm skeptical of the distinction that some people make between punching up vs down, and I don't think it's something that generative AI should be perpetuating (though I suspect, as others have said, that it comes from norms found in the training data, rather than special rules / hard-coded protections). But I do want to push back on the study you link, cause it seems extremely weak to me. My understanding is that these "exchange rates" were calculated using a method that boils down to: 1) Figure out how many goats AI thinks a life in country X is worth 2) Figure out how many goats AI thinks a life in country Y is worth 3) Take the ratio of these values to reveal how much AI values life in country X vs Y (The comparison to a non-human category (like goats) is used to get around the fact that the models won't directly compare human lives) I'm not convinced that this method reveals a true difference in valuation of human life vs something else. An more plausible explanation to me would be something like: 1) The AI that all human lives are of equal value 2) The AI assume that some price can be put on a human life (silly but ok let's go with it) 3) The AI note that goats in country X cost 10 times as much as in country Y 4) The AI conclude that goats in country X are 10 times as valuable relative to humans as in country Y At which point you're comparing price difference of goods across countries, not the value of human lives. Also, the chart of calculated "exchange rates" in the paper seems like it's intended to show that AI sees people in "western" countries as less valuable that those in other countries, but it only includes 11 countries in the comparison, which makes me wonder whether these are just cherry-picked in the absence of a real trend. | ||