You're looking at Less Wrong's discussion board. This includes all posts, including those that haven't been promoted to the front page yet. For more information, see About Less Wrong.

ikrase comments on "Stupid" questions thread - Less Wrong Discussion

40 Post author: gothgirl420666 13 July 2013 02:42AM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (850)

You are viewing a single comment's thread.

Comment author: ikrase 13 July 2013 08:22:51AM 9 points [-]
  • What's with the ems? People who are into ems seem to make a lot of assumptions about what ems are like and seem completely unattached to present-day culture or even structure of life, seem willing to spam duplicates of people around, etc. I know that Hanson thinks that 1. ems will not be robbed of their humanity and 2. that lots of things we currently consider horrible will come to pass and be accepted, but it's rather strange just how as soon as people say 'em' (as opposed to any other form of uploading) everything gets weird. Does anthropics come into it?

  • Why the huge focus on fully paternalistic Friendly AI rather than Obedient AI? It seems like a much lower-risk project. (and yes, I'm aware of the need for Friendliness in Obedient AI.)

Comment author: cousin_it 13 July 2013 08:48:17AM *  15 points [-]

For what it's worth, Eliezer's answer to your second question is here:

There is no safe wish smaller than an entire human morality. (...) With a safe genie, wishing is superfluous. Just run the genie.

Comment author: ikrase 14 July 2013 03:02:46AM 0 points [-]

I think that Obedient AI requires less fragility-of-values types of things.

Comment author: Eliezer_Yudkowsky 14 July 2013 04:22:32AM 5 points [-]

I don't see why a genie can't kill you just as hard by missing one dimension of what it meant to satisfy your wish.

Comment author: ikrase 14 July 2013 10:23:10AM 0 points [-]

I'm not talking naive obedient AI here. I'm talking a much less meta FAI that does not do analysis of metaethics or CEV or do incredibly vague, subtle wishes. (Atlantis in HPMOR may be an example of a very weak, rather irrational, poorly safeguarded Obedient AI with a very, very strange command set.)

Comment author: timtyler 14 July 2013 11:46:38AM -1 points [-]

There is no safe wish smaller than an entire human morality.

Is that true? Why can't the wish point at what it wants (e.g. the wishes of particular human X) - rather than spelling it out in detail?

Comment author: ESRogs 16 July 2013 01:05:09AM 2 points [-]

To spell out some of the complications -- does the genie only respond to verbal commands? What if the human is temporarily angry at someone and an internal part of their brain wishes them harm. The genie needs to know not to act on this. So it must have some kind of requirement for reflective equilibrium.

Suppose the human is duped into pursuing some unwise course of action? The genie needs to reject their new wishes. But the human should still be able to have their morality evolve over time.

So you still need a complete CV Extrapolator. But maybe that's what you had in mind be pointing at the wishes of a particular human?

Comment author: drethelin 14 July 2013 04:50:27PM 2 points [-]

The first problem is the wish would have to be extremely good at pointing.

This sounds silly but what I mean is that humans are COMPLICATED. "Pointing" at a human and telling an AI to deduce things about it will come up with HUGE swathes of data which you have to have already prepared it to ignore or pay attention to. To give a classic simple example, smiles are a sign of happiness but we do not want to tile the universe in smiley faces or create an artificial virus that constricts your face into a rictus and is highly contagious.

Second: assuming that works, it works primarily for one person, which is giving that person a lot more power than I think most people want to give any one person. But if we could guarantee an AI would fulfill the values of A person rather than of multiple people and someone else was developing AI that wasn't guarunteed to fulfill any values I'd probably take it.

Comment author: drethelin 13 July 2013 01:09:54PM 9 points [-]

Basically it's a matter of natural selection. Given a starting population of EMs, if some are unwilling to be copied, the ones that are willing to be copied will dominate the population in short order. If EMs are useful for work, eg valuable, then the more valuable ones will be copied more often. At that point, EMs that are willing to be copied and do slave labor effectively for no complaints will become the most copied, and the population of ems will end up being composed largely of copies of the person/people who are 1) ok with being copied, 2) ok with being modified to work more effectively.

Comment author: Eliezer_Yudkowsky 14 July 2013 04:15:54AM 7 points [-]

Well, no offense, but I'm not sure you are aware of the need for Friendliness in Obedient AI, or rather, just how much F you need in a genie.

If you were to actually figure out how to build a genie you would have figured it out by trying to build a CEV-class AI, intending to tackle all those challenges, tackling all those challenges, having pretty good solutions to all of those challenges, not trusting those solutions quite enough, and temporarily retreating to a mere genie which had ALL of the safety measures one would intuitively imagine necessary for a CEV-class independently-acting unchecked AI, to the best grade you could currently implement them. Anyone who thought they could skip the hard parts of CEV-class FAI by just building a genie instead, would die like a squirrel under a lawnmower. For reasons they didn't even understand because they hadn't become engaged with that part of the problem.

I'm not certain that this must happen in reality. The problem might have much kinder qualities than I anticipate in the sense of mistakes naturally showing up early enough and blatantly enough for corner-cutters to spot them. But it's how things are looking as a default after becoming engaged with the problems of CEV-class AI. The same problems show up in proposed 'genies' too, it's just that the genie-proposers don't realize it.

Comment author: ikrase 14 July 2013 10:30:51AM 0 points [-]

I'm... not sure what you mean by this. And I wouldn't be against putting a whole CEV-ish human morality in an AI, either. My point is that there seems to be a big space between your Outcome Pump fail example and highly paternalistic AIs of the sort that caused Failed Utopia 4-2.

It reminds me a little bit about how modern computers are only occasionally used for computation.

Comment author: Eliezer_Yudkowsky 14 July 2013 11:18:49PM 6 points [-]

Anything smarter-than-human should be regarded as containing unimaginably huge forces held in check only by the balanced internal structure of those forces, since there is nothing which could resist them if unleashed. The degree of 'obedience' makes very little difference to this fact, which must be dealt with before you can go on to anything else.

Comment author: NancyLebovitz 14 July 2013 01:24:53PM 5 points [-]

As I understand it, a AI is expected to make huge, inventive efforts to fulfill its orders as it understands them.

You know how sometimes people cause havoc while meaning well? Imagine something immensely more powerful and probably less clueful making the same mistake.

Comment author: ChristianKl 13 July 2013 01:37:48PM 4 points [-]

I know that Hanson thinks that 1. ems will not be robbed of their humanity

I don't know whether Hanson has a concret concept of 'humanity'.

Comment author: ChristianKl 13 July 2013 12:09:09PM 2 points [-]

Why the huge focus on fully paternalistic Friendly AI rather than Obedient AI? It seems like a much lower-risk project. (and yes, I'm aware of the need for Friendliness in Obedient AI.)

Because the AI is better at estimating the consequences of following an order than the person giving the order.

There also the issue that the AI is likely to act in a way that changes the order that the person gives if it's own utility criteria are about fulfilling orders.

Comment author: bogdanb 13 July 2013 06:11:48PM 1 point [-]

Also, even assuming a “right” way of making obedient FAI is found (for example, one that warns you if you’re asking for something that might bite you in the ass later), there remains the problem of who is allowed to give orders to the AI. Power corrupts, etc.

Comment author: DanielLC 13 July 2013 08:46:42PM 3 points [-]

(as opposed to any other form of uploading)

I didn't know "em" was a specific form of uploading. What form is it, and what other forms are there?

Comment author: hairyfigment 15 July 2013 02:57:39AM 2 points [-]

Besides Eliezer's rather strong-looking argument, ethically creating Obedient AI would require solving the following scary problems:

  • A "nonperson predicate" that can ensure the AI doesn't create simulations which themselves count as people. If we fail to solve this one, then I could be a simulation the AI made in order to test how people like me react to torture.

  • A way to ensure the AI itself does not count as a person, so that we don't feel sad if it eventually switches itself off. See here for a fuller explanation of why this matters.

Now, I think Wei Dai suggested we start by building a "philosophical" AI that could solve such problems for us. I don't think philosophy is a natural class. (A 'correct way to do philosophy' sounds like a fully general correct way to think and act.) But if we get the AI's goals right, then maybe it could start out restricted by flawed and overcautious answers to these questions, but find us some better answers. Maybe.

Comment author: ikrase 15 July 2013 08:00:16PM 1 point [-]

I am aware of the need for those things (part of what I mean by (need for friendliness in OAI) but as far as I can tell, Paternalistic FAI requires you to solve those problems, plus simple 'not being very powerful but insane', plus basic understandings of what matters to humans, plus incredibly meta human values matters. An OAI can leave off the last one of those problems.

Comment author: hairyfigment 15 July 2013 08:38:56PM -1 points [-]
  1. I meant that by going meta we might not have to solve them fully.

  2. All the problems you list sound nearly identical to me. In particular, "what matters to humans" sounds more vague but just as meta. If it includes enough details to actually reassure me, you could just tell the AI, "Do that." Presumably what matters to us would include 'the ability to affect our environment, eg by giving orders.' What do you mean by "very powerful but insane"? I want to parse that as 'intelligent in the sense of having accurate models that allow it to shape the future, but not programmed to do what matters to humans.'

Comment author: ikrase 16 July 2013 07:51:48AM 0 points [-]

"very powerful but insane" : AI's response to orders seem to make less than no sense, yet AI is still able to do damage. "What matters to humans": Things like the Outcome Pump example, where any child would know that not dying is supposed to be part of "out of the building", but not including the problems that we are bad at solving, such as fun theory and the like.

Comment author: Tenoke 13 July 2013 08:38:44AM *  0 points [-]

What's with the ems?

We can make more soild predictions about ems than we can about strong AI since there are less black swans regarding ems to mess up our calculations.

Does anthropics come into it?

No.

Why the huge focus on fully paternalistic Friendly AI rather than Obedient AI? It seems like a much lower-risk project.

the need for Friendliness in Obedient AI.