You're looking at Less Wrong's discussion board. This includes all posts, including those that haven't been promoted to the front page yet. For more information, see About Less Wrong.

Dorikka comments on BOOK DRAFT: 'Ethics and Superintelligence' (part 1) - Less Wrong Discussion

11 Post author: lukeprog 13 February 2011 10:09AM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (107)

You are viewing a single comment's thread. Show more comments above.

Comment author: Dorikka 14 February 2011 06:09:46AM *  1 point [-]

Extrapolated humanity decides that the best possible outcome is to become the Affront. Now, if the FAI put everyone in a separate VR and tricked him into believing that he was acting all Affront-like, then everything would be great -- everyone would be content. However, people don't just want the experience of being the Affront -- everyone agrees that they want to be truly interacting with other sentiences which will often feel the brunt of each other's coercive action.

Comment author: Eliezer_Yudkowsky 14 February 2011 06:40:23AM 3 points [-]

Original version of grandparent contained, before I deleted it, "Besides the usual 'Eating babies is wrong, what if CEV outputs eating babies, therefore a better solution is CEV plus code that outlaws eating babies.'"

Comment author: lukeprog 14 February 2011 06:20:09AM *  2 points [-]

Dorikka,

I don't understand this. If the singleton's utility function was written such that it's highest value was for humans to become the Affront, then making it the case that humans believed they were the Affront while not being the Affront would not satisfy the utility function. So why would the singleton do such a thing?

Comment author: Dorikka 15 February 2011 02:45:39AM 2 points [-]

I don't think that my brain was working optimally at 1am last night.

My first point was that our CEV might decide to go Baby-Eater, and so the FAI should treat the caring-about-the-real-world-state part of its utility function as a mere preference (like chocolate ice cream), and pop humanity into a nicely designed VR (though I didn't have the precision of thought necessary to put it into such language). However, it's pretty absurd for us to be telling our CEV what to do, considering that they'll have much more information than we do and much more refined thinking processes. I actually don't think that our Last Judge should do anything more than watch for coding errors (as in, we forgot to remove known psychological biases when creating the CEV).

My second point was that the FAI should also slip us into a VR if we desire a world-state in which we defect from each other (with similar results as in the prisoner's dilemma). However, the counterargument from point 1 also applies to this point.

Comment author: nazgulnarsil 16 February 2011 01:21:39AM 2 points [-]

I have never understood what is wrong with the amnesia-holodecking scenario. (is there a proper name for this?)

Comment author: Dorikka 16 February 2011 02:20:56AM 3 points [-]

If you want to, say, stop people from starving to death, would you be satisfied with being plopped on a holodeck with images of non-starving people? If so, then your stop-people-from-starving-to-death desire is not a desire to optimize reality into a smaller set of possible world-states, but simply a desire to have a set of sensations so that you believe starvation does not exist. The two are really different.

If you don't understand what I'm saying, the first two paragraphs of this comment might explain it better.

Comment author: nazgulnarsil 16 February 2011 02:25:44AM *  0 points [-]

thanks for clarifying. I guess I'm evil. It's a good thing to know about oneself.

Comment author: Dorikka 16 February 2011 02:30:23AM 0 points [-]

Uh, that was a joke, right?

Comment author: nazgulnarsil 16 February 2011 06:19:09AM 0 points [-]

no.

Comment author: Dorikka 16 February 2011 11:53:37PM 0 points [-]

What definition of evil are you using? I'm having trouble understanding why (how?) you would declare yourself evil, especially evil_nazgulnarsil.

Comment author: nazgulnarsil 17 February 2011 06:06:07AM 4 points [-]

i don't care about suffering independent of my sensory perception of it causing me distress.

Comment author: Dorikka 17 February 2011 03:31:49PM 0 points [-]

Oh. In that case, it might be more precise to say that your utility function does not assign positive or negative utility to the suffering of others (if I'm interpreting your statement correctly). However, I'm curious about whether this statement holds true for you at extremes, so here's a hypothetical.

I'm going to assume that you like ice cream. If you don't like any sort of ice cream, substitute in a certain quantity of your favorite cookie. If you could get a scoop of ice cream (or a cookie) for free at the cost of a million babies thumbs cut off, would you take the ice cream/cookie?

If not, then you assign a non-zero utility to others suffering, so it might be true that you care very little, but it's not true that you don't care at all.

Comment author: nazgulnarsil 18 February 2011 07:33:48AM 5 points [-]

I think you misunderstand slightly. Sensory experience includes having the idea communicated to me that my action is causing suffering. I assign negative utility to other's suffering in real life because the thought of such suffering is unpleasant.

Comment author: Sniffnoy 16 February 2011 09:03:04AM 0 points [-]

Well, it's essentially equivalent to wireheading.

Comment author: nazgulnarsil 16 February 2011 10:16:45AM 0 points [-]

which I also plan to do if everything goes tits-up.