pjeby comments on Post Your Utility Function - Less Wrong

28 Post author: taw 04 June 2009 05:05AM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (273)

You are viewing a single comment's thread. Show more comments above.

Comment author: pjeby 05 June 2009 10:21:08PM -1 points [-]

Don't take it lightly, it's a well-vetted and well-understood position, extensively discussed and agreed upon.

It's extensively discussed and agreed upon, that that is how we (for certain definitions of "we") would like it to be, and it certainly has desirable properties for say, building Friendly AI, or any AI that doesn't wirehead. And it is certainly a property of the human brain that it orients its preferences towards what it believes is the outside world - again, it has good consequences for preventing wireheading.

But that doesn't make it actually true, just useful.

It's also pretty well established as a tenet of e.g., General Semantics, that the "outside world" is unknowable, since all we can ever consciously perceive is our map. The whole point of discussing biases is that our maps are systematically biased -- and this includes our preferences, which are being applied to our biased views of the world, rather than the actual world.

I am being descriptive here, not prescriptive. When we say we prefer a certain set of things to actually be true, we can only mean that we want the world to not dispute a certain map, because otherwise we are making the supernaturalist error of assuming that a thing could be true independent of the components that make it so.

To put it another way, if I say, "I prefer that the wings of this plane not fall off", I am speaking about the map, since "wings" do not exist in the territory.

IOW, our statements about reality are about the intersection of some portion of "observable" reality and our particular mapping (division and labeling) of it. And it cannot be otherwise, since to even talk about it, we have to carve up and label the "reality" we are discussing.

Comment author: loqi 07 June 2009 07:12:56AM *  0 points [-]

IOW, our statements about reality are about the intersection of some portion of "observable" reality and our particular mapping (division and labeling) of it. And it cannot be otherwise, since to even talk about it, we have to carve up and label the "reality" we are discussing.

It's funny that you talk of wordplay a few comments back, as it seems that you're the one making a technically-correct-but-not-practically-meaningful argument here.

If I may attempt to explore your position: Suppose someone claims a preference for "blue skies". The wirehead version of this that you endorse is "I prefer experiences that include the perception I label 'blue sky'". The "anti-wirehead" version you seem to be arguing against is "I prefer actual states of the world where the sky is actually blue".

You seem to be saying that since the preference is really about the experience of blue skies, it makes no sense to talk about the sky actually being blue. Chasing after external definitions involving photons and atmospheric scattering is beside the point, because the actual preference wasn't formed in terms of them.

This becomes another example of the general rule that it's impossible to form preferences directly about reality, because "reality" is just another label on our subjective map.

As far as specifics go, I think the point you make is sound: Most (all?) of our preferences can't just be about the territory, because they're phrased in terms of things that themselves don't exist in the territory, but at best simply point at the slice of experience labeled "the territory".

That said, I think this perspective grossly downplays the practical importance of that label. It has very distinct subjective features connecting in special ways to other important concepts. For the non-solipsists among us, perhaps the most important role it plays is establishing a connection between our subjective reality and someone else's. We have reason to believe that it mediates experiences we label as "physical interactions" in a manner causally unaffected by our state of mind alone.

When I say "I prefer the galaxy not to be tiled by paperclips", I understand that, technically, the only building blocks I have for that preference are labeled experiences and concepts that aren't themselves the "stuff" of their referents. In fact, I freely admit that I'm not exactly sure what constitutes "the galaxy", but the preference I just expressed actually contains a massive number of implicit references to other concepts that I consider causally connected to it via my "external reality" label. What's more, most people I communicate with can easily access a seemingly similar set of connections to their "external reality" label, assuming they don't talk themselves out of it.

The territory concept plays a similar role to that of an opaque reference in a programming language. Its state may not be invariant, but its identity is. I don't have to know any true facts concerning its actual structure for it to be meaningful and useful. Just as photons aren't explicitly required to subjectively perceive a blue sky, the ontological status of my territory concept doesn't really change its meaning or importance, which is acquired through its intimate connection to massive amounts of raw experience.

Claiming my preferences about the territory are really just about my map is true in the narrow technical sense that it's impossible for me to refer directly to "reality", but doing so completely glosses over the deep, implicit connections expressed by such preferences, most primarily the connection between myself and the things I label "other consciousnesses". In contrast, the perception of these connections seems to come for free by "confusing" the invariant identity of my territory concept with the invariant "existence" of a real external world. The two notions are basically isomorphic, so where's the value in the distinction?

Comment author: Vladimir_Nesov 07 June 2009 10:12:52AM *  1 point [-]

Think of what difference is there between "referring directly" to the outside reality and "referring directly" to the brain. Not much, methinks. There is no homunculus whose hands are only so long to reach the brain, but not long enough to touch your nose.

Comment author: loqi 07 June 2009 08:21:37PM 0 points [-]

Agreed, as the brain is a physical object. Referring "directly" to subjective experiences is a different story though.

Comment author: timtyler 05 June 2009 11:00:35PM 0 points [-]

Whether your preferences refer to your state, or to the rest of the world is indeed a wirehead-related issue. The problem with the idea that they refer to your state is that that idea tends to cause wirehead behaviour - surgery on your own brain to produce the desired state. So - it seems desirable to construct agents that believe that there is a real world, and that their preferences relate to it.

Comment author: pjeby 05 June 2009 11:07:23PM -1 points [-]

Whether your preferences refer to your state, or to the rest of the world is indeed a wirehead-related issue. The problem with the idea that they refer to your state is that that idea tends to cause wirehead behaviour - surgery on your own brain to produce the desired state. So - it seems desirable to construct agents that believe that there is a real world, and that their preferences relate to it.

I agree - that's probably why humans appear to be constructed that way. The problem comes in when you expect the system to also be able to accurately reflect its preferences, as opposed to just executing them.

This does not preclude the possibility of creating systems that can; it's just that they're purely hypothetical.

To the greatest extent practical, I try to write here only about what I know about the practical effects of the hardware we actually run on today, if for no other reason than if I got into entirely-theoretical discussions I'd post WAY more than I already do. ;-)

Comment author: timtyler 05 June 2009 11:14:09PM 0 points [-]

Presumably, if you asked such an agent to reflect on its own purposes, it would claim that they related to the external world (unless it's aim was to deceive you about its purposes for signalling reasons, of course).

For example, it might claim that its aim was to save the whales - rather than to feel good about saving the whales. It could do the latter by taking drugs or via hypnotherapy - and that is not how it actually acts.

Comment author: pjeby 05 June 2009 11:28:31PM 0 points [-]

Presumably, if you asked such an agent to reflect on its own purposes, it would claim that they related to the external world (unless it's aim was to deceive you about its purposes for signalling reasons, of course).

Actually, if signaling was its true purpose, it would claim the same thing. And if it were hacked together by evolution to be convincing, it might even do so by genuinely believing that its reflections were accurate. ;-)

For example, it might claim that its aim was to save the whales - rather than to feel good about saving the whales. It could do the latter by taking drugs or via hypnotherapy - and that is not how it actually acts.

Indeed. But in the case of humans, note first that many people do in fact take drugs to feel good, and second, that we tend to dislike being deceived. When we try to imagine getting hypnotized into believing the whales are safe, we react as we would to being deceived, not as we would if we truly believed the whales were safe. It is this error in the map that gives us a degree of feed-forward consistency, in that it prevents us from certain classes of wireheading.

However, it's also a source of other errors, because in the case of self-fulfilling beliefs, it leads to erroneous conclusions about our need for the belief. For example, if you think your fear of being fired is the only thing getting you to work at all, then you will be reluctant to give up that fear, even if it's really the existence of the fear that is suppressing, say, the creativity or ambition that would replace the fear.

In each case, the error is the same: System 2 projection of the future implicitly relies on the current contents of System 1's map, and does not take into account how that map would be different in the projected future.

(This is why, by the way, The Work's fourth question is "who would you be without that thought?" The question is a trick to force System 1 to do a projection using the presupposition that the belief is already gone.)