orthonormal comments on Consequentialism Need Not Be Nearsighted - Less Wrong

53 Post author: orthonormal 02 September 2011 07:37AM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (118)

You are viewing a single comment's thread. Show more comments above.

Comment author: Jack 31 August 2011 04:51:19AM *  6 points [-]

This post really brings to light an inkling I had a while ago: TDT feels vaguely Kantian.

Compare:

  1. "Choose as though controlling the logical output of the abstract computation you implement, including the output of all other instantiations and simulations of that computation."

  2. "Act only according to that maxim whereby you can at the same time will that it should become a universal law without contradiction."

Now, they're clearly not the same but they are similar enough that we shouldn't be surprised that consequentialism under TDT alleviates some of our concerns about traditional consequentialism. I find this exciting-- but it also makes me suspicious. Kantian moral theory has some serious problems and so I wonder if there might be analogous issues in CON+TDT. And I see some. I'll leave out the Kantian equivalent unless someone is interested:

"What happens in general if everyone at least as smart as me deduces that I would do X whenever I'm in situation Y"?

The problem is that no two situations are strictly speaking identical (putting aside exact simulations and other universes). That means CON+TDT doesn't prohibit a decision to carve up a vagrant for organs conditional on some unique feature of the situation. Or put it this way: can TDT correctly model intuitions about moral scenarios that cannot plausibly be replicated?

There are lots of possible agents that closely resemble TDT agents but with slight differences that would make them difficult to model. It seems like absent complete transparency rational agents should publicly endorse and abide by TDT but deviate to the extent they can get away with it.

The relevant scenarios can be too broad as well. For example, say you think a violent response to Hitler was justified. Unless Hitler is resurrected that algorithm won't get used again (save simulations and other universes). But if we pick something broader that can be multiply instantiated we can run into trouble- "Violent response to leaders we disagree with is justified."

So is there a rigorous way of determining classes of decisions?

Comment author: orthonormal 31 August 2011 02:12:26PM 5 points [-]

You've identified a subtle problem with implementing decision theories, but the answer that the commonsense version of TDT "should" give is pretty clear: if the differences between two situations don't massively affect the utilities involved (from the perspective of the deciding agent), then they should belong to the same reference class.

If you shouldn't kill the patient in a house, then you still shouldn't kill him with a mouse, or on a boat, with a goat, on a train, in the rain, here or there, or anywhere.

Comment author: nazgulnarsil 05 September 2011 06:05:05PM 3 points [-]

this is somewhat circular. It works for the example but not knowing how similar two situations have to be before similar decisions produce similar utility is part of the problem.