▲ | caseyy a day ago | |||||||
That’s an interesting point. If we created super-intelligence but it wasn’t anthropomorphic, we might just not consider it super-intelligent as a sort of ego defence mechanism. Much good (and bad) sci-fi was written about this. In it, usually this leads to some massive conflict that forces humans to admit machines as equals or superiors. If we do develop super-intelligence or consciousness in machines, I wonder how that will all go in reality. | ||||||||
▲ | yieldcrv 19 hours ago | parent [-] | |||||||
Some things I think about are how different the goals could be For example, human and biological based goals are around self-preservation and propagation. And this in turn is about resource appropriation to facilitate that, and systems of doing that become wealth accumulation. Species that don't do this don't continue existing. A different branch of evolution of intelligence may take a different approach, that allows its affects to persist anyway. | ||||||||
|