| ▲ | paulmooreparks 5 hours ago | |
There is no "should" in the relevant section. It's making a mathematical model of the risks and benefits. > Now consider a choice between never launching superintelligence or launching it immediately, where the latter carries an % risk of immediate universal death. Developing superintelligence increases our life expectancy if and only if: > [equation I can't seem to copy] > In other words, under these conservative assumptions, developing superintelligence increases our remaining life expectancy provided that the probability of AI-induced annihilation is below 97%. | ||