cousin_it comments on "Stupid" questions thread - Less Wrong

40 Post author: gothgirl420666 13 July 2013 02:42AM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (850)

You are viewing a single comment's thread. Show more comments above.

Comment author: cousin_it 13 July 2013 08:48:17AM *  15 points [-]

For what it's worth, Eliezer's answer to your second question is here:

There is no safe wish smaller than an entire human morality. (...) With a safe genie, wishing is superfluous. Just run the genie.

Comment author: ikrase 14 July 2013 03:02:46AM 0 points [-]

I think that Obedient AI requires less fragility-of-values types of things.

Comment author: Eliezer_Yudkowsky 14 July 2013 04:22:32AM 5 points [-]

I don't see why a genie can't kill you just as hard by missing one dimension of what it meant to satisfy your wish.

Comment author: ikrase 14 July 2013 10:23:10AM 0 points [-]

I'm not talking naive obedient AI here. I'm talking a much less meta FAI that does not do analysis of metaethics or CEV or do incredibly vague, subtle wishes. (Atlantis in HPMOR may be an example of a very weak, rather irrational, poorly safeguarded Obedient AI with a very, very strange command set.)

Comment author: timtyler 14 July 2013 11:46:38AM -1 points [-]

There is no safe wish smaller than an entire human morality.

Is that true? Why can't the wish point at what it wants (e.g. the wishes of particular human X) - rather than spelling it out in detail?

Comment author: ESRogs 16 July 2013 01:05:09AM 2 points [-]

To spell out some of the complications -- does the genie only respond to verbal commands? What if the human is temporarily angry at someone and an internal part of their brain wishes them harm. The genie needs to know not to act on this. So it must have some kind of requirement for reflective equilibrium.

Suppose the human is duped into pursuing some unwise course of action? The genie needs to reject their new wishes. But the human should still be able to have their morality evolve over time.

So you still need a complete CV Extrapolator. But maybe that's what you had in mind be pointing at the wishes of a particular human?

Comment author: drethelin 14 July 2013 04:50:27PM 2 points [-]

The first problem is the wish would have to be extremely good at pointing.

This sounds silly but what I mean is that humans are COMPLICATED. "Pointing" at a human and telling an AI to deduce things about it will come up with HUGE swathes of data which you have to have already prepared it to ignore or pay attention to. To give a classic simple example, smiles are a sign of happiness but we do not want to tile the universe in smiley faces or create an artificial virus that constricts your face into a rictus and is highly contagious.

Second: assuming that works, it works primarily for one person, which is giving that person a lot more power than I think most people want to give any one person. But if we could guarantee an AI would fulfill the values of A person rather than of multiple people and someone else was developing AI that wasn't guarunteed to fulfill any values I'd probably take it.