mattnewport comments on Open Thread: May 2009 - Less Wrong

4 Post author: steven0461 01 May 2009 04:16PM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (204)

You are viewing a single comment's thread. Show more comments above.

Comment author: mattnewport 07 May 2009 12:30:09AM *  0 points [-]

I agree that the 11th man has no moral obligation in this scenario and I wouldn't consider him to have acted immorally to choose his own life. I think this is the kind of scenario where people will signal altruism by costlessly saying they would sacrifice themselves in the thought experiment but that far fewer people would sacrifice themselves when faced with the actual choice.

I suspect that people will tend to look negatively on those who fail to signal altruism by saying they would sacrifice themselves but would be more forgiving of someone who was faced with the choice in reality. I think they would consider an individual who had actually made such a choice to have made a morally weak but understandable choice and would not consider him deserving of punishment.

Comment author: conchis 11 May 2009 12:58:18PM *  4 points [-]

You two seem to be making slightly different points here. Matt, I take it you accept that there is some reason to sacrifice yourself (not doing so would be "morally weak") but that failing to do so would not be blameworthy. That sounds like a fairly mainstream view. In contrast, MrHen seems to be making the stronger claim that there is no reason to save the others at all (unless he has a personal investment in said others).

The idea that [the King is responsible for the deaths] screens off the possibility that [the 11th man is responsible for the deaths] seems to be a version of the single-true-cause fallacy. Sure, the king is responsible, but given the king's actions, it's the 11th man's choice that directly determines whether the others will live or not.

If you want to prioritize your own life over theirs then you are free to do so, but I think you should own up to the fact that that's ultimately what you're doing. Disclaiming responsibility entirely seems like a convenient excuse designed to let you get what you want without having to feel bad about it.

Comment author: MrHen 12 May 2009 03:42:56AM 0 points [-]

I have to read the single-true-cause fallacy before I can fully reply, but here is a quick ditty to munch on until then:

Sure, the king is responsible, but given the king's actions, it's the 11th man's choice that directly determines whether the others will live or not.

I disagree with this. The eleventh's choice is completely irrelevant. The king has a decision to make and just because he makes it the same every single time does not mean the actual decision is different the next time around.

The similar example where the king puts a gun in the eleventh's hand and says "kill them or I kill you" is when the choice actually becomes the eleventh's. In this scenario, the eleventh man has to choose to (a) kill the ten or (b) not kill the ten. This is a moral decision.

Of note, whoever actually has to kill the ten has this choice and will probably choose the selfish route. If the king shares the blame with anyone, it will be whoever actually kills the ten. If the eleventh is morally responsible than everyone else watching the event is morally responsible, too.

Comment author: conchis 12 May 2009 10:19:15AM 1 point [-]

The eleventh's choice is completely irrelevant.

I don't understand what coherent theory of causation could make this statement true.

If the eleventh is morally responsible than everyone else watching the event is morally responsible, too.

If they could stop it, then yes, they are.

Comment author: MrHen 12 May 2009 01:19:56PM *  0 points [-]

I don't understand what coherent theory of causation could make this statement true.

The issue is not causality. The issue is moral responsibility. If I go postal and start shooting people as they run past my house and later tell the police that it was because my neighbor pissed me off, the neighbor may have been one (of many) causes but should not be held morally responsible for my actions.

Likewise, if the king asks someone a question and, in response, kills ten people, I do not see how the question asks makes any different in the assignment of moral responsibility.

Causality does not imply moral responsibility.

Also, having read the link you gave earlier, I can now comment on this:

The idea that [the King is responsible for the deaths] screens off the possibility that [the 11th man is responsible for the deaths] seems to be a version of the single-true-cause fallacy. Sure, the king is responsible, but given the king's actions, it's the 11th man's choice that directly determines whether the others will live or not.

"Responsible" has two meanings. The first is a cause-effect sense of "these actions precluded these other actions." This is the same as saying a bowling ball is responsible for the bowling pins falling over.

The other is a moral judgement stating "this person should be held accountable for this evil." The bowling ball holds no moral responsibility because it was thrown by a bowler.

I am not claiming that the eleventh man was not part of the causal chain that resulted in ten people dying. I am claiming that the eleventh man holds no moral responsibility for the ten people dying. I am not trying to say that the king is the single-true-cause. I am claiming that the king is the one who should be held morally responsible.

To belabor this point with one more example: If I rigged a door to blow up when opened and Jack opened the door while standing next to Jill they are both reduced to goo. Jack is causally responsible for what happened because he opened the door. He is not, however, morally responsible.

The question of when someone does become morally responsible is tricky and I do not have a good example of when I think the line is crossed. I do not, however, pass any blame on the eleventh man for answer a question to which there is no correct answer.

Comment author: conchis 12 May 2009 03:18:53PM *  0 points [-]

The issue is not causality. The issue is moral responsibility.

Agreed. But I think if you want to separate the two, you need a reasonable account of the distinction. One plausible account relies on reasonably foreseeable consequences to ground responsibility, and this is pretty much my view. It accounts easily for the neighbor, bowling ball, and Jack and Jill cases, but still implies responsibility for the 11th man.

I can accept a view that says that, all things considered, the king has a greater causal influence on the outcome of the 11th man case, and thus bears much greater moral responsibility for it than does the 11th man. But (and this was the point of the no-single-true-cause analogy) I see no reason why this should imply that the 11th man has no responsibility whatsoever, given that the death of 10 innocent others is a clearly foreseeable consequence of his choice.

I still think this is a convenient conclusion designed to let you be selfish without feeling like you're doing anything wrong.

P.S. FWIW, yes I pretty much do think you're evil if you're not willing to sacrifice $100 to save 10 lives in your hostage example. I can understand not being willing to die, even if I think it would be morally better to sacrifice oneself. (And I readily confess that it's possible that I would take the morally wrong/weak choice if actually faced with this situation.) But for $100 I wouldn't hesitate.

Comment author: MrHen 12 May 2009 03:44:34PM 1 point [-]

One plausible account relies on reasonably foreseeable consequences to ground responsibility, and this is pretty much my view.

I can understand that. I have not dug quite so deeply into this area of my ethical map so it could be representing the territory poorly. What little mental exercises I have done have led me to this point.

I guess the example that really puts me in a pickle is asking what would happen if Jack knew the door was rigged but opened it anyway. It makes sense that Jack shares the blame. There seems to be something in me that says the physical action weighs against Jack.

So, if I had to write it up quickly:

  • Being a physical cause in a chain of events that leads to harm
  • While knowing the physical action has a high likelihood of leading to harm
  • Is evil

But, on the other hand:

  • Being a non-physical cause in a chain of events that leads to harm
  • While knowing the non-physical action has a high likelihood of leading to harm
  • Is not necessarily evil but can be sometimes

Weird. That sure seems like an inconsistency to me. Looks like I need to get the mapmaking tools out. The stickiness of the eleventh man is that the king is another moral entity and the king somehow shrouds the eleventh from actually making a moral choice. But I do not have justification for that distinction.

There may yet be justification, but working backwards is not proper. Once I get the whole thing worked out I will report what I find, if you are interested.

Comment author: conchis 12 May 2009 05:10:38PM 0 points [-]

Good luck with the map-making! I'd certainly be interested to know what you find, if and when you find it.

Comment author: mattnewport 11 May 2009 05:43:08PM *  0 points [-]

My use of the phrase 'morally weak' was to describe how I think many/most people would view the choice, not my own personal judgement. I agree with MrHen that the 11th man's choice is not morally wrong. I was contrasting that with what I think would be the mainstream view that the choice is morally wrong but understandable and not deserving of punishment.

To me this is similar to the trolley problems where you are supposed to choose between taking action and killing one person to save 10 or taking no action and allowing the 10 to die. The one person to be sacrificed is yourself however. I wouldn't kill the one to save the 10 either (although I view that as more morally wrong than sacrificing yourself). I also generally place much lower moral weight on harm caused by inaction than harm caused by action and the forced choice scenario here presents the 11th man with a situation that I think is similar to one of causing harm by inaction.

Comment author: conchis 11 May 2009 08:10:49PM 0 points [-]

Sorry, my bad. Thanks for clearing that up.

As to the act-omission distinction, it would be simple enough to stipulate that the default option is that you die unless you tell the king to kill the other ten. Does this change your willingness to die?

Comment author: mattnewport 11 May 2009 08:20:43PM 0 points [-]

No, that wouldn't change my decision. It's the not-sacrificing-your-life that I'm comparing with causing harm by inaction (the inaction being the not-sacrificing) rather than anything specific about the way the question is phrased.

The agency of the king does make a relevant difference in this scenario in my view. It is not exactly equivalent to a scenario where you could sacrifice your life to save 10 people from a fire or car crash. Although I don't think there is a moral obligation in that case either I do consider the difference morally relevant.

Comment author: JGWeissman 11 May 2009 09:57:32PM 1 point [-]

Suppose the king has 10 people prepared to be hung. They are in the gallows with nooses around their neck, standing on a trap door. The king shows you a lever that will open the trap door, and kill the 10 victims. The king informs you that if you do not pull the lever within one hour, the 10 people will be freed and you will be executed.

Here the king has set up the situation, but you will be the last sentient being capable of moral reasoning in the causal chain that kills 10 people. Is your conclusion different in this scenario?

Comment author: mattnewport 11 May 2009 10:23:52PM 0 points [-]

The king here is more diabolical and the scenario you describe is more traumatic. I believe it does change the intuitive moral response to the scenario. I don't believe it changes my conclusion of the morality of the act. I feel that I'd still direct my moral outrage at the king and absolve the 11th man of moral responsibility.

This is where these kinds of artificial moral thought experiments start to break down though. In real situations analogous to this I believe the uncertainty in the outcomes of various actions (together with other unspecified details of the situation) would overwhelm the 'pure' decision made on the basis of the thought experiment. I'm unconvinced of the value of such intuition pumps in enhancing understanding of a problem.

Comment author: conchis 11 May 2009 11:22:20PM *  1 point [-]

Why is this where the thought experiments suddenly start to break down? Sure, it's a less convenient world for you, but I don't see why it's any more artificial than the original problem, and you didn't seem to take issue with that.

Comment author: mattnewport 11 May 2009 11:58:24PM 0 points [-]

I have taken issue with the use of thought experiments generally in previous comments, partly because it seems to me that they start to break down rapidly when pushed further into 'least convenient world' territory. I'm skeptical in general of the value of thought experiments in revealing philosophical truths of any kind, ethical or otherwise. They are often designed by construction to trigger intuitive judgements based on scenarios so far from actual experience that those judgements are rendered highly untrustworthy.

I answered the original question to say that yes, I did agree that the 11th man was not acting immorally here. I suspect this particular thought experiment is constructed as an intuition pump to generate the opposite conclusion and to the extent that the first commenter is correct that the view that the 11th man has done nothing immoral is a minority position it would seem it serves its purpose.

I've attempted to explain why I think the intuition that this is morally questionable is generated and why I think it's not to be fully trusted. I don't intend to endorse the use of such thought experiments as a good method for examining moral questions though.

Comment author: conchis 12 May 2009 11:44:48AM *  1 point [-]

Fair enough. It was mainly the appearance of motivated stopping that I was concerned with.

While I share some general concerns about the reliability of thought experiments, in the absence of a better alternative, the question doesn't seem to be whether we use them or not, but how we can make best use of them despite their potential flaws.

In order to answer that question, it seems like we might need a better theory of when they're especially likely to be poor guides than we currently have. It's not obvious, for example, that their information content increases monotonically in realism. Many real-world issues seem too complicated for simple intuitions to be much of a guide to anything.*

As well as trying to frame scenarios in ways that reduce noise/bias in our intuitions, we can also try to correct for the effect of known biases. A good example would be adjusting for scope insensitivty. But we need to be careful about coming up with just-so stories to explain away intuitions we disagree with. E.g. you claim that the altruist intuition is merely a low cost-signal; I claim that the converse is merely self-serving rationalization. Both of these seem like potentially good examples of confirmation bias at work.

Finally, it's worth bearing in mind that, to the extent that our main concern is that thought experiments provide noisy (rather than biased) data, this could suggest that the solution is more thought experiments rather than fewer (for standard statistical reasons).

* And even if information content did increase with realism, realism doesn't seem to correspond in any simple way to convenience (as your comments seem to imply). Not least because convenience is a function of one's favourite theory as much as it is a function of the postulated scenario.

Comment author: dclayh 12 May 2009 01:43:39AM 0 points [-]

They are often designed by construction to trigger intuitive judgments based on scenarios so far from actual experience that those judgments are rendered highly untrustworthy.

Agreed; however it's important to distinguish between this sort of appeal-to-intuition and the more rigorous sort of thought experiment that appeals to reasoning (e.g. Einstein's famous Gedankenexperimente).

Comment author: JGWeissman 11 May 2009 10:56:48PM 1 point [-]

I don't believe it changes my conclusion of the morality of the act.

Given that your defense of the morality was based on the inaction of not self sacrificing, and that in this scenario inaction means self sacrifice and you have to actively kill the other 10 people to avoid it, what reasoning supports keeping the same conclusion?

Comment author: mattnewport 11 May 2009 11:17:33PM 0 points [-]

I'm comparing the inaction to the not-self-sacrificing, not to the lack of action. I attempted to clarify the distinction when I said the similarity was not 'anything specific about the way the question is phrased'.

The similarity is not about the causality but about the cost paid. In many 'morality of inaction' problems the cost to self is usually so low as to be neglected but in fact all actions carry a cost. I see the problem not as primarily one of determining causality but more as a cost-benefit analysis. Inaction is usually the 'zero-cost' option, action carries a cost (which may be very small, like pressing a button, or extremely large, like jumping in front of a moving trolley). The benefit is conferred directly on other parties and indirectly on yourself according to what value you place on the welfare of others (and possibly according to other criteria).

I think our moral intuition is primed to distinguish between freely chosen actions taken to benefit ourselves that ignore fairly direct negative consequences on others (which we generally view as morally wrong) and refraining from taking actions that would harm ourselves but would fairly directly benefit others (which may or may not be viewed as morally wrong but are generally seen as 'less wrong' than the former). We also seem primed to associate direct action with agency and free choice (since that is usually what it represents) and so directly taken actions tend to lead to events being viewed as the former rather than the latter.

I believe the moral 'dilemma' represented by carefully constructed thought experiments like this represents a conflict between our 'agency recognizing' intuition that attempts to distinguish directly taken action from inaction and our judgement of sins of commission vs. omission. Given that the unusual part of the dilemma is the forced choice imposed by a third party (the evil king) it seems likely that the moral intuition that is primed to react to agency is more likely to be making flawed judgements.

Comment author: conchis 12 May 2009 12:00:56PM 0 points [-]

I see the problem not as primarily one of determining causality but more as a cost-benefit analysis.

This makes sense to me, but it seems to run counter to the nature of MrHen's original claim that the issue is lack of responsibility. For example, if it's all about CBA, then you would presumably be more uneasy about MrHen's hostage example ($100 vs. 10 lives) than he seems to be. Presumably also you would become even more uneasy were it $10, or $1, whereas MrHen's argument seems to suggest that all of this is irrelevant because you're not responsible either way.

Am I understanding you correctly?

Comment author: conchis 11 May 2009 09:46:20PM *  0 points [-]

OK, I think I'm sort of with you now, but I'm just want to be clear about the nature of the similarity claim you're making. Is it that:

  1. you think there's some sort of justificatory similarity between not-sacrificing and harm-by-inaction such that you those who are inclined to allow harm-by-inaction, should therefore also be more willing to allow not-sacrificing; or is it just that
  2. you just happen to hold both the view that harm-by-inaction is allowed and the view that not-sacrificing is allowed, but the justifications for these views are independent (i.e. it's merely a contingent surface similarity)?

I originally assumed you were claiming something along the lines of 1. but I'm struggling to see how such a link is supposed to work, so maybe I've misinterpreted you're intention.

Comment author: mattnewport 11 May 2009 10:12:49PM 0 points [-]

you think there's some sort of justificatory similarity between not-sacrificing and harm-by-inaction such that you those who are inclined to allow harm-by-inaction, should therefore also be more willing to allow not-sacrificing

Yes. I'd generally hold that it is not morally wrong to allow harm-by-inaction: there is not a general moral obligation to act to prevent harm. In real moral dilemmas there is a continuum of cost to the harm-preventing action and when that cost is low relative to the harm prevented it would be morally good to perform that action but not morally required. At extremely low cost relative to harm things become a little fuzzy and inaction borders on an immoral choice. When the cost of the action is extremely high (likely or certain self-sacrifice) then there is no fuzziness and inaction is clearly morally allowed (not-sacrificing by jumping in front of a trolley cart to save 10 is not immoral).

Given inaction being morally permitted in the trolley case, I have difficulty imagining a coherent moral system that would then say that it was not permissible for the 11th man to save himself. The evil king does change the problem but I can only see it making not-sacrificing more rather than less morally acceptable. I can conceive of coherent moral systems that would allow the 11th man to save himself but would require the trolley jumper to sacrifice himself. I have difficulty conceiving of the reverse. That's not to say that one doesn't exist, it's just sufficiently removed from my own moral sense that it doesn't present itself to me.

Comment author: steven0461 11 May 2009 10:17:02PM 0 points [-]

This thought experiment among other things convinces me that omission vs. commission is a sliding scale.

Comment author: mattnewport 11 May 2009 10:28:09PM *  0 points [-]

That would fall in the territory I describe as fuzzy above. At a sufficiently low cost inaction begins to seem morally questionable. That is largely driven by intuition though and I'm skeptical of attempts to scale it up and draw moral conclusions. I believe there are reasons the intuition exists that do not scale up simply. In other words, scaling up from this to conclude that if a very small cost is obligatory to save a single person then a very large cost is obligatory to save a million people is faulty reasoning in my opinion.