Alicorn comments on Post Your Utility Function - Less Wrong

28 Post author: taw 04 June 2009 05:05AM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (273)

You are viewing a single comment's thread. Show more comments above.

Comment author: Alicorn 09 June 2009 06:56:41AM 0 points [-]

I wish I could upvote this two or three times. Thank you.

Comment author: Vladimir_Nesov 09 June 2009 12:19:16PM 0 points [-]

What features of that comment made it communicate something new to you? What was it that got communicated?

The comment restated a claim that a certain relationship is desirable as a claim that given that it's desirable, there is a process that establishes it to be true. It's interesting how this restatement could pierce inferential distance: is preference less trustworthy than a fact, and so demonstrating the conversion of preference into a fact strengthens the case?

Comment author: Alicorn 09 June 2009 04:22:12PM *  0 points [-]

I'd been following this topic and getting frustrated with my inability to put my opinion on the whole preferences-about-the-territory thing into words, and I thought that orthonomal's comment accomplished it very nicely. I don't think I understand your other question.

Comment author: orthonormal 09 June 2009 11:50:00PM 0 points [-]

Given the length of the thread I branched from, it looks like you and P.J. Eby ended up talking past each other to some extent, and I think that you both failed to distinguish explicitly between the current map (which is what you calculate the territory to be) and a hypothetical future map.

P.J. Eby was (correctly) insisting that your utility function is only in contact with your current map, not the territory directly. You were (correctly) insisting that your utility function cares about (what it calculates to be) the future territory, and not just the future map.

Is that a fair statement of the key points?

Comment author: Vladimir_Nesov 10 June 2009 12:36:56AM 0 points [-]

Utility function is no more "in contact" with your current map than the actual truth of 2+2=4 is in contact with display of a calculator that displays the statement. Utility function may care about past territory (and even counterfactual territory) as well as future territory, with map being its part. Keeping a map in good health is instrumentally a very strong move: just by injecting an agent with your preferences somewhere in the territory you improve it immensely.

Comment author: orthonormal 10 June 2009 03:38:12PM 0 points [-]

While there might exist some abstracted idealized dynamic that is a mathematical object independent of your map, any feasible heuristic for calculating your utility function (including, of course, any calculation you actually do) will depend on your map.

If Omega came through tomorrow and made all pigs conscious with human-like thoughts and emotions, my moral views on pig farming wouldn't be instantly changed; only when information about this development gets to me and my map gets altered will I start assigning a much higher disutility to factory farming of pigs.

Or, to put it another way, a decision algorithm refers directly to the possible worlds in the territory (and their probabilities, etc), but it evaluates these referents by looking at the corresponding objects in its current map. I think that, since we're talking about practical purposes, this is a relevant point.

Keeping a map in good health is instrumentally a very strong move: just by injecting an agent with your preferences somewhere in the territory you improve it immensely.

Agree completely. Of the worlds where my future map looks to diverge from the territory, though, I'm generally more repulsed by the ones in which my map says it's fine where it's not than by the opposite.

Comment author: conchis 10 June 2009 03:51:37PM *  1 point [-]

any feasible heuristic for calculating your utility function (including, of course, any calculation you actually do) will depend on your map.

This something of a nitpick, but this isn't strictly true. If others are trying to calculate your utility function (in order to help you), this will depend on their maps rather than yours (though probably including their map of your map). The difference becomes important if their maps are more accurate than yours in some respect (or if they can affect how accurate your map is).

For example, if you know that I value not being deceived (and not merely the subjective experience of not being deceived), and you care about my welfare, then I think that you should not deceive me, even if you know that I might perceive my welfare to be higher if you did.

Comment author: orthonormal 10 June 2009 03:55:06PM 0 points [-]

Oh, good point. I should have restricted it to "any calculation you personally do", in which case I believe it holds.

Comment author: Vladimir_Nesov 10 June 2009 04:24:28PM *  0 points [-]

At which point it becomes trivial: any calculation that is done on your map is done using your map, just Markovity of computation...

A related point is that you can create tools that make decisions themselves, in situations only of possibility of which you are aware.

Comment author: orthonormal 10 June 2009 04:34:55PM 0 points [-]

Right. It's trivial, but relevant when discussing in what sense our decision algorithms refer to territory versus map.

A related point is that you can create tools that make decisions themselves, in situations only of possibility of which you are aware.

I can't parse this. What do you mean?

Comment author: Vladimir_Nesov 10 June 2009 04:46:16PM 0 points [-]

If you install an alarm system that uses a video camera to recognize movement and calls the police if it's armed, you are delegating some of the map-making and decision-making to the alarm system. You are neither aware of the exact nature of possible intruders, nor making a decision regarding calling the police before any intrusion actually occurs. The system decides what to do by itself, according to the aspect of your values it implements. You map is not involved.