Eliezer_Yudkowsky comments on Domesticating reduced impact AIs - Less Wrong

9 Post author: Stuart_Armstrong 14 February 2013 04:59PM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (104)

You are viewing a single comment's thread. Show more comments above.

Comment author: Eliezer_Yudkowsky 18 February 2013 05:55:39PM 11 points [-]

I got 99 psychological drives but inclusive fitness ain't one.

In what way is evolution supposed to be robust? It's slow, stupid, doesn't reproduce the content of goal systems at all and breaks as soon as you introduce it to a context sufficiently different from the environment of evolutionary ancestry because it uses no abstract reasoning in its consequentialism. It is the opposite of robust along just about every desirable dimension.

Comment author: Stuart_Armstrong 19 February 2013 11:35:09AM *  4 points [-]

In what way is evolution supposed to be robust?

It's not as brittle as methods like first order logic or computer programming. If I had really bad computer hardware (corrupted disks and all that), then an evolved algorithm is going to work a lot better than a lean formal program.

Similarly, if an AI was built by people who didn't understand the concept of friendliness, I'd much prefer they used reinforcement learning or evolutionary algorithms than direct programming. With the first approaches, there is some chance the AI may infer the correct values. But with the wrong direct programming, there's no chance of it being safe.

As you said, you're altruistic, even if the laws of physics change - and yet you don't have a full theory of humankind, of worth, of altruism, etc... So the mess in your genes, culture and brain has come up with something robust to ontology changes, without having to be explicit about it all. Even though evolution is not achieving it's "goal" through you, something messy is working.

Comment author: CCC 19 February 2013 08:13:37AM 2 points [-]

In what way is evolution supposed to be robust?

If I had to guess Stuart_Armstrong's meaning, I would guess that genetic algorithms are robust in that they can find a solution to a poorly specified and poorly understood problem statement. They're not robust to dramatic changes in the environment (though they can correct for sufficiently slow, gradual changes very well); but their consequentialist nature provides some layer of protection from ontology crises.

Comment author: Evan_Crowe 18 February 2013 09:05:57PM 2 points [-]

I think there might be a miscommunication going on here.

I see Stuart arguing that Genetic algorithms function independent of physics in terms of their consistent "friendly" trait. i.e. if in universe a there is a genetic algorithm that finds value in expressing the "friendly" trait, then that algorithm would, if placed in universe b (where the boundary conditions of the universe were slightly different) would tend to eventually express that "friendly" trait again. Thereby meaning robust (when compared to systems that could not do this)

I don't necessarily agree with that argument, and my interpretation could be wrong.

I see Eliezer arguing that evolution as a system doesn't do a heck of a lot, when compared to a system that is designed around a goal and involves compensation for failure. i.e. I can't reproduce with a horse, this is a bad thing because if I were trapped on an island with a horse our genetic information would die off, where in a robust system, I could breed with a horse, thereby preserving our genetic information.

I'm sorry if this touches too closely on the entire "well, the dictionary says" argument.

Comment author: Evan_Crowe 18 February 2013 10:57:20PM 0 points [-]

Oh, now I feel silly. The horse IS the other universe.

Comment author: Stuart_Armstrong 19 February 2013 11:36:45AM *  0 points [-]

Comment author: MugaSofer 19 February 2013 02:51:32PM *  0 points [-]

You know this is blank, right?

Comment author: Stuart_Armstrong 19 February 2013 03:01:37PM 2 points [-]

I had a response that was mainly a minor nitpick; it didn't add anything, so I removed it.

Comment author: MugaSofer 19 February 2013 03:13:54PM -2 points [-]

Just checking.