Vladimir_Nesov comments on Harry Potter and the Methods of Rationality discussion thread, part 5 - Less Wrong

6 Post author: NihilCredo 02 November 2010 06:57PM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (648)

You are viewing a single comment's thread. Show more comments above.

Comment author: Vladimir_Nesov 07 November 2010 09:39:19PM *  3 points [-]

Bella's life isn't just less valuable than Harry's. Her life has a large negative value. Harry should be trying to prevent her escape.

If she kills again, then yes, she shouldn't be saved. But this is a claim about consequences unrelated to the value of her life itself. If she can be saved and prevented from killing others, she should be.

Comment author: Eugine_Nier 08 November 2010 05:26:54AM 1 point [-]

By TDT, the fact that she did a lot of evil things in the past is reason enough to assign her life less value.

Comment author: nshepperd 08 November 2010 10:34:32AM 2 points [-]

Only if how much value people like Harry and Quirrel would assign to her life was a significant factor in her decisionmaking. Chances are that at the time the idea that someone might want to save her from azkaban never even crossed her mind.

Comment author: Nick_Tarleton 08 November 2010 05:34:16AM 2 points [-]

For game-theoretic reasons? (If so, even though this isn't about consequences, it still seems very worth separating from the terminal value of her life.)

Comment author: Eugine_Nier 08 November 2010 05:59:35AM 3 points [-]

(If so, even though this isn't about consequences, it still seems very worth separating from the terminal value of her life.)

I'm not sure what you mean by this. Part of my point is that even if "she can be saved and prevented from killing others", she probably still shouldn't be.

Comment author: Vladimir_Nesov 08 November 2010 08:58:38AM 4 points [-]

It's still about consequences, just not causal consequences. It might be that saving the murderer has an acausal consequence of leading to more murders in the past. This should of course be taken into account, and is harder to guarantee.

Comment author: DanArmak 08 November 2010 02:28:50PM *  4 points [-]

If you buy the acausal "reality is not turing-computable" view of the HPMOR universe, then none of our usual reasoning methods work all that well. I'm surprised Harry hasn't been making a bigger deal of this fact.

Comment author: David_Gerard 13 November 2010 08:58:38PM *  1 point [-]

"It might be that saving the murderer has an acausal consequence of leading to more murders in the past."

... This sentence makes no sense to me at all as a statement in conventional English.

What are the meanings of the words in this sentence, what are the beliefs that I would need to hold for this sentence to make sense, and how would they pay their rent?

(I am not asking for a list of sequences to be directed to. I am asking for the translation from deep LessWrong jargon into conventional English.)

Comment author: Alicorn 13 November 2010 09:26:56PM 9 points [-]

I'll try my hand at the translation, although the rent-paying thing is above my pay grade.

Something like: "If you save a murderer, that means you are the sort of person disposed to save a murderer. That means that in the past, murderers, insofar as they had accurate beliefs about your dispositions, will have been less disincentivized to murder."

Comment author: Perplexed 13 November 2010 10:40:34PM 13 points [-]

what are the beliefs that I would need to hold for this sentence [about acausal consequences] to make sense, and how would they pay their rent?

the rent-paying thing is above my pay grade.

Above my pay grade too, but as I am an amateur, I won't let that deter me.

First, you would need to believe that free will is an oversimplification. More specifically, that what may appear to be a free-will moral decision made today (about saving a murderer, say) is actually a decision the making of which is spread over your entire past life (for example, the point in your life where you formed moral opinions about murder, revenge, and so on). And not just spread over your life, but actually spread over the entire history of our species, in the course of which the genes and cultural traditions that contribute to your own moral intuitions were formed.

Second, you would have to believe that your moral decision today is so correlated to those aspects of the past, and those aspects are so correlated in a causal and deterrent way to the past behavior of potential murderers, that your decision nominally made today about punishing a murderer is correlated "kinda-sorta-causally" with the number of past murders.

And third, you would have to realize that if you actually refer to this as a "causal" relationship, you probably will no longer be invited to the best dinner parties, and therefore you would choose to call this relationship "acausal". It is basically a way of signaling your own mental hygiene - scare quotes would also suffice, but the word acausal has become entrenched.

Rent paying. Hmmm. I'm going to get cute here and say that this holding these beliefs is not not really about paying your rent. It is about paying your taxes. Your duty to society and all that.

Comment author: TheOtherDave 15 November 2010 06:24:29PM 1 point [-]

Um?

OK, I would say that your first belief implies that what appears to be a decision in this case is in fact not a decision, but rather a working out of the inevitable consequences of an earlier state. The only reason it seems like a decision is because I'm ignorant of the real possibilities.

The situation, on this account, is analogous to my "deciding" not to fly away when dropped off the top of a building. I can imagine being in a delusional state where it does not seem inevitable to me that I will fall to the ground, and I can further imagine being in a state where I believe I have decided to fall, but In both cases I would simply be wrong.

Similarly, on this account, my belief that I choose whether to save a murderer or not is simply wrong. There's no actual decision to be made; that there seems to be a decision is simply a delusion shared by approximately everyone.

It's like having an electrical switch connected to both a buzzer and a lightbulb, with a 3-second delay between the switch being flipped and the lightbulb going on. Clearly any intuition that the buzzer caused the light to turn on is just post hoc ergo prompter hoc run amok, but to start talking about the lightbulb having an acausal consequence of the buzzer going off seems downright unjustified.

It is not clear to me why that situation changes if the buzzer is a murder and the lightbulb is me saving the murderer, or if the buzzer is Omega putting money into boxes and the lightbulb is me opening the box. Sure, the underlying mechanisms are way more complex, but I don't see how any of that complexity affects any of what we're talking about, if we start from the assumption that free will is an illusion.

(shrug) Maybe the whole point here is to resurrect some notion of free will worth considering, or something... I dunno. I am still working my way fairly linearly through the old OB posts, and have not actually read any of the more recent TDT stuff, so probably I shouldn't be trying to engage in these conversations just yet.

But, well, sometimes I do things I shouldn't.

Comment author: TobyBartels 18 November 2010 08:28:53AM 2 points [-]

I would say that your first belief implies that what appears to be a decision in this case is in fact not a decision, but rather a working out of the inevitable consequences of an earlier state.

Do you ever make a decision that is not like this?

I think that the official Less Wrong answer to the problem of free will is that you do make a decision, since it is the consequence of your state, but this is just as a computer may make a decision, say to allocate certain CPU cycles to certain processes (which is the sort of decision that modern operating system kernels are designed to make, and one that my computer is not making very wisely at the moment, which is why I thought of it). Given the input, this decision is inevitable, but it's arguably still a decision.

Comment author: TheOtherDave 18 November 2010 04:15:16PM 2 points [-]

For what it's worth, I'm a compatibilist as well, although I don't think it's a particularly important question.

I'd merely meant to point out that if it's possible (as stipulated in this example) to predict accurately at T1 what I'm going to do at T2, then there's no new salient information added to the system after T1, so it's as reasonable to talk about Omega's behavior at T1 being determined by the state of the world at T1 as it is to talk about it as being determined retroactively by the state of the world at T2.

(Perplexed has since then pointed out that the second formulation is simpler in some sense, and therefore potentially useful, which I accept.)

That being said... as Perplexed articulates well here, it's hard to understand the purpose of decision theory in the first place from a compatibilist or determinist stance.

Comment author: saturn 15 November 2010 07:37:12PM 3 points [-]

In a situation where other people can predict your future decisions, you can reason as if your decisions have backward causality, because they were predicted before you carried them out. It's a generalization of the concept of making credible threats and promises.

Comment author: TheOtherDave 15 November 2010 08:13:26PM 1 point [-]

But in a situation where other people can predict my future behavior, I can instead reason as if some earlier state causes both my behavior and their prediction.

This seems to get me all the same explanatory and predictive power in an entirely straightforward fashion. Making credible threats and promises seems entirely unproblematic when looked at that way.

I accept that some awfully smart people who have thought about this a lot have concluded that this sort of backward-causality reasoning really does buy them something, and I'm open to the possibility that it's something worth buying. But at the moment I don't see what it could possibly be.

Comment author: Perplexed 15 November 2010 08:35:01PM 2 points [-]

if we start from the assumption that free will is an illusion...

Nominally, decision theory is all about giving good advice to people who make decisions.

Now I am willing to entertain the idea that the free will of the decision maker is an illusion.

But my 'willing suspension of disbelief' goes all to hell when I ask myself: "Why does an illusion need good advice?"

Decision theory absolutely requires an assumption that the will is free in some sense. However, it does seem reasonable to consider the possibility that free decision making can be spread out in time.

Traditional game theory assumes that an agent freely chooses a set of preferences over states-of-the-world well in advance. Then, at decision time, he chooses an action so as to maximize the probability of reaching a desirable state of the world. Classical game theory and decision theory offer advice on that second free decision, but they don't advise on those earlier free decisions which created the preference schedule. Perhaps they should.

Or, perhaps we need an additional theory, over and above game theory and decision theory, which will advise agents on how to set their preferences so as to take into account some of the side effects of those preferences. What do we call this new kind of normative theory? 'Moral theory', perhaps?

Comment author: TheOtherDave 15 November 2010 10:04:28PM 2 points [-]

Um. Let me taboo some words ("free will," "prediction", "decision") here and try again.

Let us suppose that at time T1 someone either commits murder (event E1a) or doesn't (E1b), and at T2 I either spare the murderer (E2a) or don't (E2b). (I don't mean to suggest here that all combinations are possible.)

The original scenario seemed to presuppose that at T1 there is a fact of the matter about whether, given E1a, T2 contains E2a or E2b, and that some potential murderers are able to use that fact in their reasoning.

My understanding is that some people are saying we can therefore understand E2a|b to have some kind of "acausal" influence on E1a|b. (If that's not true, then I've utterly misunderstood either the scenario or the conversation or both, which is entirely plausible.)

I agree with you that it's useful to talk about what happens at T1 (or earlier)... that is, to advise about preference schedules. Indeed, as you suggest, it's hard to see what the point of advising about anything else in this scenario would be.

But I don't understand why it isn't just as useful, to this end, to say that there exists at T1 some set of facts S describing the state of the world (including my mind), and that E2 a|b and E1 a|b both depend on S, as it is to say that E2a|b exerts an acausal atemporal influence on E1 a|b.

What does that second formulation buy us?

Comment author: thomblake 03 May 2012 06:51:18PM 0 points [-]

I like this explanation of what "acausal" means.

Comment author: TobyBartels 15 November 2010 09:14:32AM 0 points [-]

Wow, I think that I finally understand TDT (assuming that you accurately described it). Thanks!

Comment author: Perplexed 15 November 2010 05:00:08PM 1 point [-]

Wow, I think that I finally understand TDT

Cool. Maybe you can explain it to me, then. ;)

I'm pretty sure that my sketch doesn't capture all of TDT. Only a rationalization of the "acausal influence" aspect of it.

Comment author: NihilCredo 15 November 2010 08:05:23PM 4 points [-]

The only practical application of TDT that I've found is as an anti-procrastination argument. "If this excuse seems good enough to me, it's probably going to seem about as good to me+15 minutes, so if I don't get to work now I'm very unlikely to do so in 15 minutes either". I still find the existence of significant acausal connections between different individuals to be pretty much claptrap.

Comment author: Kutta 16 November 2010 10:32:44AM *  3 points [-]

My short take: your decision algorithm that outputs saving or not saving the murderer is instantiated multiple times. Anyone who tries to predict your output also runs a more or less precise simulation of your algorithm. Suppose a perfect predictor murderer in the past. In this case, no matter what your decision is, the prediction was the same.

So, you can reason this way: "although I don't know my final decision yet, I know that it correlates with the prediction perfectly. Therefore I also have to consider the consequences and resulting utilities of the prediction when making the decision. Shouldn't I just act then as if was controlling the output of both my current algorithm and that of the predictor, weighing the utilities together? I should output a decision now such that maximizes utility over present and past, because the past prediction mirrors the current me perfectly."

And if there are imperfect predictors involved (or algorithms with imperfectly correlated outputs), you reason as if you had imperfect control over their outputs. As far as I managed to understand it, this is TDT. Note that there is some interesting self-referentiality: the TDT algorithm computes the expected utility of its own "possible" outputs, and then makes output with maximum utility.

Comment author: steven0461 08 November 2010 05:53:55AM *  1 point [-]

Even if she were completely reformed with no chance of killing more people, the terminal value of her life could still be lower than that of a random person. For example, presumably continuity is valuable (or we wouldn't mind replacing dead people with random newborns), and there's less continuity between evil and good than between good and good; and someone with an evil past will probably experience less happiness during sincere reflection on that past (for some morally relevant value of "sincere") than someone with a good past. It also seems possible that there are components of terminal value that track the game-theoretic reasons you mention: we may terminally value game-theoretically-defined notions of "cooperation" or "justice" on top of their instrumental usefulness.

Comment author: PhilGoetz 07 November 2010 09:47:30PM 0 points [-]

If she kills again, it's too late. I don't understand how you can divorce the value of her life from the consequences of her life.

Comment author: Vladimir_Nesov 07 November 2010 09:50:48PM *  3 points [-]

Hence the conditional in the last sentence. And it's never "too late", the same reasoning applies no matter how many people she murders.

Comment author: PhilGoetz 08 November 2010 05:01:54AM *  0 points [-]

It is too late to save the person that she murders.

Given that she's already murdered many people, why would her murdering one more change your decision?

If she can be saved and prevented from killing others, she should be.

Why?

Is this an absolute moral position, without regard for how she is prevented from killing others? I thought you frowned on those. If she can be saved and prevented from killing others by locking her in a cell for the rest of her life, she should be? If that's not the plan, what is?

Comment author: thomblake 08 November 2010 06:37:55PM 3 points [-]

If that's not the plan, what is?

The plan as stated is to bring her to a magic shrink who will undo the mental torture that made her this way, so she stops being evil.

Comment author: wnoise 08 November 2010 06:48:11PM 2 points [-]

Of course, that's just killing her and replacing her with someone similar.

Comment author: bentarm 16 November 2010 01:50:25AM 8 points [-]

So did you just kill me and replace me with a version of me that had read that post? Where do we draw the line?

Comment author: wnoise 16 November 2010 04:36:31PM 0 points [-]

It's often hard to draw a clear line.

It's quite true that there's a continuum of personality modifications between "talking to someone" and "using magic to eradicate emotions and base personality features replacing them with more convenient ones". But even without drawing the line specifically, it's clear that these two examples are on different sides of that line.

Comment author: bentarm 17 November 2010 01:53:10AM 0 points [-]

Well, cryonics aside, it's actually relatively easy for most people to draw a clear line between "killed" and "not-killed", and both of those examples seem to fall quite squarely on the "not-killed" side, as they don't result in death.

You clearly draw your line somewhere very different to most people. You should probably at least attempt to explain what you mean by "kill" if you're going to use it in such a non-standard way.

Comment author: NihilCredo 16 November 2010 01:46:28AM 5 points [-]

Of course, if that process is considered a form of killing, one should stop and re-elaborate their opinion of "killing" in light of this expanded definition before passing judgment.

Comment author: wnoise 16 November 2010 04:39:03PM *  2 points [-]

Sure. I never said it wasn't a justified killing. This Bellatrix "needs killing", and if we can come close enough to restoring a previous mentality that doesn't, that's a great bonus.

Assuming, of course, that all of this is as Quirrel said.

Comment author: TheOtherDave 16 November 2010 05:37:57PM 4 points [-]

Sure. But NihilCredo's point is actually broader than that.

When I adopt an importantly different new referent for a term, it's really not a good idea for me to carry forward old ideas associated with that term without re-evaluating them. It's a special case of the common problem of confusing the label for the referent.

In some cases, the confusion is obvious and easy to avoid. For example, a couple of years ago the U.S. adopted a new referent for the term "current President of the United States" that made previously true statements (e.g., "the current President of the United States is a member of the Republican party") suddenly false. This doesn't really confuse anyone; we understand that the referent for that label changes over time.

But "killing" isn't like that for most people. If I adopt an understanding of "killing" congruent with your comment above I ought to stop and ask myself things like "Is killing a bad thing? Why? Is there any reason I shouldn't go around killing everyone I can? What reason? Is there any reason I shouldn't kill myself every opportunity I get?" and so forth. But there's a strong chance I won't.

At some point it becomes less confusing to just adopt a new word.

In other words, the issue here is not whether magical therapy on Bellatrix is a justified killing or an unjustified killing, but whether it's useful or consistent to refer to it as any kind of killing at all.

Completely tangential to this: I have a lot of difficulty reconciling the idea that reliable psychiatric intervention to "heal" people of criminal tendencies exists in the HPverse with the idea that Azkaban exists there. It's as if a fictional world contained the ability to cure any disease, but wealthy and important people in that world were forced to die of diseases without being cured. I don't claim it's impossible -- human societies are perfectly capable of contradictions this absurd -- but it would be awfully difficult to defend.

Comment author: Dan074 16 November 2010 05:45:49PM 3 points [-]

Azkaban already seems difficult to defend, versus just killing people in the ordinary sense. Sure, killing people is inhumane, but much less inhumane, and much more secure, than imprisoning people for life in a place that destroys all happiness and goodness.

Comment author: wnoise 17 November 2010 06:49:02AM 1 point [-]

but whether it's useful or consistent to refer to it as any kind of killing at all.

I was raised with fantasy and sci-fi. For me "death" has always meant "death-of-personality". Short of that, who cares? And if that's gone, who cares if the body is still there?

If I adopt an understanding of "killing" congruent with your comment above I ought to stop and ask myself things like "Is killing a bad thing?

Yeah, you should. And you should say "generally, yes". For exactly the same reasons that your current understanding of killing is that it's generally bad -- it's the destruction of a person.