Will_Newsome comments on A Rationalist's Tale - Less Wrong
You are viewing a comment permalink. View the original post to see all comments and the full post content.
You are viewing a comment permalink. View the original post to see all comments and the full post content.
Comments (305)
My problem with Will's outlook is that if we are indeed being "watched over by a superintelligence", it doesn't appear to care about us in any very helpful way. Our relationship to it is therefore more about survival than it is about morality. According to the scenario, there is some thing out there which is all-powerful, whose actions depend partly on our actions, and which doesn't care about {long list of evolutionary and historical holocausts}, in any way that we would recognize as caring. Clearly, if we had any idea of the relationship between our actions and its actions, it would be in one's interest, first of all, to act so that it would not allow various awful things to happen to you and anyone you care about, and second, to act so that you might gain some advantage from its powers.
It appears that the only distinctive reason Will has for entertaining such a scenario is the usual malarkey about timeless game-theoretic equilibria... A while back, I was contemplating a post, to be called "Towards a critique of acausal reason", which was going to mention three fallacies of timeless decision theory: acausal democracy, acausal trade, acausal blackmail. The last two arise from a fallacy of selective attention: to believe them possible, you must only pay attention to possible worlds which only care about you in a highly specific way. But for any possible world where there is an intelligence simulating your response and which will do X if you do Y, there is another possible world where there is an intelligence which will do X if you don't do Y. And the actual multiplicity of worlds in which intelligences make decisions on the basis of decisions made by agents in other possible worlds that they are simulating it is vanishingly small, in the set of all possible worlds. Why the hell would you base your decision, regarding what to do in your own reality, on the opinions or actions of a possible entity in another world? You may as well just flip a coin. The whole idea that intelligences in causally disjoint worlds are in a position to trade, bargain, or arrive at game-theoretic equilibria is deeply flawed; it's only a highly eccentric agent which "cares" strongly about events which are influenced by only an extremely small fraction of its subjective duplicates (its other selves in the space of possible worlds). So some of these "eccentric agents" may genuinely "do deals", but there is no reason to think that they are anything more than a vanishingly small minority among the total population of the multiverse. (Obviously it would be desirable for people trying to work rigorously in TDT to make this argument in a rigorous form, but I don't see anything that's going to change the basic conclusion.)
So that leaves us in the more familiar situation, of possibly being in a simulation, or possibly facing the rise of a superintelligence in the near future, or possibly being somewhere in the guts of a cosmic superintelligence which either just tolerates our existence because we haven't crossed thresholds-of-caring yet, or which has a purpose for us which extends to tolerating the holocausts I mentioned earlier. All of this suggests that our survival and well-being are on the line, but it doesn't suggest that we are embedded in an order that is moral in any conventional sense.