gRR comments on Holden's Objection 1: Friendliness is dangerous - LessWrong

11 Post author: PhilGoetz 18 May 2012 12:48AM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (428)

You are viewing a single comment's thread. Show more comments above.

Comment author: gRR 19 May 2012 03:44:52AM *  -1 points [-]

Yes, it is only a hypothesis. Until we actually built an AI with such CEV as utility, we cannot know whether it could function. But at least, running it is uncontroversial by definition.

And I think I'll be more surprised if anyone was found who really and truly had a terminal value for universal death. With some strain, I can imagine someone preferring it conditionally, but certainly not absolutely. The members of doomsday cults, I expect, are either misinformed, insincere, or unhappy about something else (which FAI could fix!).

Comment author: DanArmak 19 May 2012 03:35:16PM *  0 points [-]

Until we actually built an AI with such CEV as utility, we cannot know whether it could function. But at least, running it is uncontroversial by definition.

It's quite controversial. Supposing CEV worked exactly as expected, I still wouldn't want it to be done. Neither do some others in this thread. And I'm sure neither would most humans in the street if you were to ask them (and they seriously though about the question).

CEV doesn't and cannot predict that the extrapolated wishes of everybody will perfectly coincide. Rather, it says it will find the best possible compromise. Of course I would prefer my own values to a compromise! Lacking that, I would prefer a compromise over a smaller group whose members were more similar to myself (such as the group of people actually building the AI).

I might choose CEV over something else because plenty of other things are even worse. But CEV is very very far from the best possible thing, or even the best not-totally-implausible AGI I might expect in my actual future.

And I think I'll be more surprised if anyone was found who really and truly had a terminal value for universal death

Any true believer in a better afterlife qualifies: there are billions of people who at least profess such beliefs, so I expect some of them really believe.

Comment author: gRR 19 May 2012 04:34:44PM 1 point [-]

CEV doesn't and cannot predict that the extrapolated wishes of everybody will perfectly coincide. Rather, it says it will find the best possible compromise.

What I proposed in this thread is that CEV would forcibly implement only the (extrapolated) wish(es) of literally everyone. Regarding the rest, it is to minimize its influence, leaving all decisions to people.

Any true believer in a better afterlife qualifies

No, because they believe in afterlife. They do not wish for universal death. Extrapolating their wish with correct knowledge solves the problem.

Comment author: DanArmak 19 May 2012 05:02:41PM 0 points [-]

What I proposed in this thread is that CEV would forcibly implement only the (extrapolated) wish(es) of literally everyone.

Well then, as I and others argue elsewhere in the thread, we anticipate there will be no extrapolated wishes that literally everyone agrees on.

(And that's even without considering some meta formulations of CEV that propose to also take into account the wishes of counterfactual people who might exist in the future, and dead ones who existed in the past.)

No, because they believe in afterlife. They do not wish for universal death. Extrapolating their wish with correct knowledge solves the problem.

Lots of people religiously believe that their god has planned (and prophesied) a specific event of drastic universal change, after which future people will stop suffering in this world, or will stop being born to a life of negative utility (end of the world), or will be rescued from horrible eternal torture (Hell), or which is necessary for the true believers to actually be resurrected or to enter the good afterlife. (Obviously people don't believe all of this at once; these are variant examples.)

Some others believe that life in this world is suffering, negative utility, and ought to be stopped for its own sake (stopping the cycle of rebirth).

Comment author: gRR 19 May 2012 05:10:58PM 0 points [-]

we anticipate there will be no extrapolated wishes that literally everyone agrees on

Well, now you know there exist people who believe that there are some universally acceptable wishes. Let's do the Aumann update :)

Lots of people religiously believe...

False beliefs => irrelevant after extrapolation.

Some others believe that life in this world is suffering, negative utility, and ought to be stopped for its own sake (stopping the cycle of rebirth)

False beliefs (rebirth, existence of nirvana state) => irrelevant after extrapolation.

Comment author: DanArmak 19 May 2012 05:29:58PM 0 points [-]

Well, now you know there exist people who believe that there are some universally acceptable wishes. Let's do the Aumann update :)

Aumann update works only if I believe you're a perfect Bayesian rationalist. So, no thanks.

Since you aren't giving any valid examples of universally acceptable wishes (I've pointed out people who don't wish for the examples you gave), why do you believe such wishes exist?

False beliefs => irrelevant after extrapolation.

Only if you modify these actual people to have their extrapolated beliefs instead of their current ones. Otherwise the false current beliefs will keep on being very relevant to them. Do you want to do that?

Comment author: gRR 19 May 2012 05:41:32PM 0 points [-]

Aumann update works only if I believe you're a perfect Bayesian rationalist. So, no thanks.

Too bad. Let's just agree to disagree then, until the brain scanning technology is sufficiently advanced.

I've pointed out people who don't wish for the examples you gave

So far, I didn't see a convincing example of a person who truly wished for everyone to die, even in extrapolation.

Otherwise the false current beliefs will keep on being very relevant to them

To them, yes, but not to their CEV.

Comment author: DanArmak 19 May 2012 06:05:58PM 0 points [-]

Too bad. Let's just agree to disagree then, until the brain scanning technology is sufficiently advanced.

Or until you provide the evidence that causes you to hold your opinions.

So far, I didn't see a convincing example of a person who truly wished for everyone to die, even in extrapolation.

I think it's plausible such people exist. Conversely, if you fine-tune your implementation of "extrapolation" to make their extrapolated values radically different from their current values (and incidentally matching your own current values), that's not what CEV is supposed to be about. But before talking about that, there's a more important point:

To them, yes, but not to their CEV.

So why do you care about their extrapolated values? If you think CEV will extrapolate something that matches your current values but not those of many others; and you don't want to change by force others' actual values to match their extrapolated ones, so they will suffer in the CEV future; then why extrapolate their values at all? Why not just ignore them and extrapolate your own, if you have the first-mover advantage?

Comment author: gRR 19 May 2012 06:26:39PM -1 points [-]

why extrapolate values at all

Extrapolated values are the true values. Whereas the current values are approximations, sometimes very bad and corrupted approximations.

they will suffer in the CEV future

This does not follow.

Comment author: DanArmak 19 May 2012 06:51:00PM 1 point [-]

Extrapolated values are the true values. Whereas the current values are approximations, sometimes very bad and corrupted approximations.

What makes you give them such a label as "true"? There is no such thing as a "correct" or "objective" value. Or values are possible in the sense that there can be agents will all possible values, even paperclip-maximizing. The only interesting property of values is who actually holds them. But nobody actually holds your extrapolated values (today).

Current values (and values in general) are not approximations of any other values. All values just are. Why do you call them approximations?

they will suffer in the CEV future

This does not follow.

In your CEV future, the extrapolated values are maximized. Conflicting values, like the actual values held today by many or all people, are necessarily not maximized. In proportion to how much this happens, which is positively correlated to the difference between actual and extrapolated values, people who hold the actual values will suffer living in such a world. (If the AI is a singleton they will not even have a hope of a better future.)

Briefly: suffering ~ failing to achieve your values.

Comment author: Dolores1984 19 May 2012 05:18:32AM 0 points [-]

But at least, running it is uncontroversial by definition.

I'm very dubious of CEV as a model for Friendly AI. I think it's a bad idea for several reasons. So, not that either.

Also, on topic, recall that, when you extrapolate the volition of crazy people, their volition is not, in particular, more sane. It is more as they would like to be. If you see lizard people, you don't want to see lizard people less. You want sharpened senses to detect them better. Likewise, if you extrapolate a serial killer, you don't get Ghandi. You get an incredibly good serial killer.

Comment author: gRR 19 May 2012 11:32:53AM 1 point [-]

I'm very dubious of CEV as a model for Friendly AI. I think it's a bad idea for several reasons. So, not that either.

I don't see how this is possible. One can be dubious about whether it can be defined in the way it is stated, or whether it can be implemented. But assuming it can, why would it be controversial to fulfill the wish(es) of literally everyone, while affecting everything else the least?

when you extrapolate the volition of crazy people, their volition is not, in particular, more sane

Extrapolating volition includes correcting wrong knowledge and increasing intelligence. So, you do stop seeing lizard people if they don't exist.

Serial killers are more interesting example. But they too don't want everyone to die. Assuming serial killers get full knowledge of their condition and sufficient intelligence for understanding it, what would their volition actually be? I don't know, but I'm sure it's not universal death.

Comment author: Dolores1984 19 May 2012 08:10:20PM 1 point [-]

But assuming it can, why would it be controversial to fulfill the wish(es) of literally everyone, while affecting everything else the least?

Problems:

Extrapolation is poorly defined, and, to me, seems to go in either one of two directions: either you make people more as they would like to be, which throws any ideas of coherence out the window, or you make people 'better' a long a specific axis, in which case you're no longer directing the question back at humanity in a meaningful sense. Even something as simple as removing wrong beliefs (as you imply) would automatically erase any but the very weakest theological notions. There are a lot of people in the world who would die to stop that from happening. So, yes, controversial.

Coherence, one way or another, is unlikely to exist. Humans want a bunch of different things. Smarter, better-informed humans would still want a bunch of different, conflicting things. Trying to satisfy all of them won't work. Trying to satisfy the majority at the expense of the minorities might get incredibly ugly incredibly fast. I don't have a better solution at this time, but I don't think taking some kind of vote over the sum total of humanity is going to produce any kind of coherent plan of action.

Comment author: [deleted] 19 May 2012 10:09:59PM 0 points [-]

Trying to satisfy the majority at the expense of the minorities might get incredibly ugly incredibly fast.

But would that be actually uglier than the status quo? Right now, to a very good approximation, those who were born from the right vagina are satisfied at the expense of those born from the wrong vagina. Is that any better?

I call the Litany of Gendlin on the idea that everyone can't be fully satisfied at once. And I also call the Fallacy of Gray on the idea that if you can't do something perfectly, then doing it decently is no better than not doing it at all.

Comment author: Dolores1984 20 May 2012 12:06:19AM 0 points [-]

But would that be actually uglier than the status quo?

I don't know. It conceivably could be, and there would be no possibility of improving it, ever. I'm just saying it might be wise to have a better model before we commit to something for eternity.

Comment author: gRR 19 May 2012 09:23:02PM 0 points [-]

For extrapolation to be conceptually plausible, I imagine "knowledge" and "intelligence level" to be independent variables of a mind, knobs to turn. To be sure, this picture looks ridiculous. But assuming, for the sake of argument, that this picture is realizable, extrapolation appears to be definable.

Yes, many religious people wouldn't want their beliefs erased, but only because they believe them to be true. They wouldn't oppose increasing their knowledge if they knew it was true knowledge. Cases of belief in belief would be dissolved if it was known that true beliefs were better in all respects, including individual happiness.

Coherence, one way or another, is unlikely to exist. Humans want a bunch of different things...

Yes, I agree with this. But, I believe there exist wishes universal for (extrapolated) humans, among which I think there is the wish for humans to continue existing. I would like for AI to fulfil this wish (and other universal wishes if there are any), while letting people decide everything else for themselves.

Comment author: TheOtherDave 19 May 2012 02:53:32PM 1 point [-]

But assuming it can, why would it be controversial to fulfill the wish(es) of literally everyone, while affecting everything else the least?

It is not clear that CEV as a model for FAI does either of those things.

Comment author: gRR 19 May 2012 03:51:03PM 0 points [-]

AFAIK, CEV is not well-defined or fully specified, except as a declaration of intent, a research direction. Thus, it does not make sense to say whether CEV as a model for FAI does or does not in fact do specific things. It only makes sense to say whether the intention of CEV's developers for it to do or not do those things, and whether CEV's specification so far contradicts or does not contradict those things.

AFAIU, CEV's developers' intent and CEV's specification so far (with added "unanimousity" condition, if it is not present in the standard CEV specification) do not contradict my statement.

Comment author: TheOtherDave 19 May 2012 05:43:07PM 1 point [-]

Just to make sure I understand your claim: you're asserting that we can identify some set of people in the world right now who are "CEV's developers," and if we asked them "does CEV fulfill the wish(es) of literally everyone while affecting everything else the least?" they would agree that it clearly does?

Comment author: gRR 19 May 2012 05:51:26PM 0 points [-]

No, because "does CEV fulfill....?" is not a well-defined or fully specified question. But I think, if you asked "whether it is possible to build FAI+CEV in such a way that it fulfills the wish(es) of literally everyone while affecting everything else the least", they would say they do not know.

Comment author: TheOtherDave 19 May 2012 08:15:22PM 0 points [-]

Ah, OK. I completely misunderstood your claim, then. Thanks for clarifying.