Will superintelligent machines destroy humanity? Or will they suffer a super-existential meltdown? Or will they develop their own super-moral system that tends towards goodness?