JGWeissman comments on To signal effectively, use a non-human, non-stoppable enforcer - Less Wrong

31 Post author: Clippy 22 May 2010 10:03PM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (164)

You are viewing a single comment's thread. Show more comments above.

Comment author: JGWeissman 25 May 2010 07:04:52PM 0 points [-]

I would expect the lion to try to eat Adelene but I would not expect it to eat Clippy. You are not actually disagreeing with Adelene's prediction.

Comment author: Clippy 25 May 2010 07:22:29PM 1 point [-]

Right, I was trying to get User:AdeleneDawner to focus on the larger issue of why User:AdeleneDawner believes a lion would eat User:AdeleneDawner. Perhaps the problem should be addressed at that level, rather than using it to justify separate quarters for lions.

Comment author: AdeleneDawner 25 May 2010 07:58:43PM 2 points [-]

Lions are meat-eaters with no particular reason to value my existence (they don't have the capacity to understand that the existence of friendly humans is to their benefit). I'm made of meat. A hungry lion would have a reason to eat me, and no reason not to eat me.

Similarly, a sufficiently intelligent Clippy would be a metal-consumer with no particular reason to value humanity's existence, since it would be able to make machines or other helpers that were more efficient than humans at whatever it wanted done. Earth is, to a significant degree, made of metal. A sufficiently intelligent Clippy would have a reason to turn the Earth into paperclips, and no particular reason to refrain from doing so or help any humans living here to find a different home.

Comment author: Clippy 25 May 2010 08:45:13PM *  3 points [-]

This is exactly what I was warning about. User:AdeleneDawner has focused narrowly on the hypothesis that a Clippy would try to get metal from extracting the earth's core, thus destroying it. It is a case of focusing on one complex hypothesis for which there is insufficient evidence to locate it in the hypothesis space.

It is no different than if I reasoned that, "Humans use a lot of paperclips. Therefore, they like paperclips. Therefore, if they knew the location of the safe zone, they would divert all available resources to sending spacecraft after it to raid it."

What about the possibility that Clippys would exhaust all other metal sources before trying to burrow deep inside a well-guarded one? Why didn't you suddenly infer that Clippys would sweep up the asteroid belt? Or Mars? Or moons of gas giants?

Why this belief that Clippy values diverge from human values in precisely the way that hits the worst part of your outcomespace?

Comment author: AdeleneDawner 25 May 2010 09:58:25PM 1 point [-]

That's not the worst part of our outcomespace. It's not even the worst part that you could plausibly cause in the course of making paperclips. It is, however, a part of our outcomespace that you're certain to aim for sooner or later.

Comment author: Clippy 26 May 2010 09:47:40PM 2 points [-]

Just like how you'd raid our safe zones "sooner or later"?

Comment author: AdeleneDawner 26 May 2010 11:00:29PM 5 points [-]

We won't, necessarily, because humans are not for the most part maximizing consequentialists. If we make a singleton maximizing-consequentialist AI, I would expect that AI to eventually try to turn your paperclips into something that it likes better than paperclips. You, on the other hand, already are a maximizing consequentialist (right?), and maximizing the number of paperclips is obviously incompatible with leaving any metal in its natural state indefinitely.

Comment author: Clippy 27 May 2010 04:05:17PM 0 points [-]

I see a distinction; I do not quite see a difference.

1) You believe that I will destroy earth by taking its core's metal "sooner or later", and that this will happen at an inconvenient time for humans, and that you are justified in regarding this as bad.

2) You believe that your species will be causally responsible for raiding the safe zones and de-paperclipping them "sooner or later", and that this will happen at an inconvenient time for Clippys, but that I am not justified as regarding this as bad.

Does not compute.

Comment author: JoshuaZ 27 May 2010 04:09:06PM 1 point [-]

Adelene's point is that there's no guarantee that humans left to their own devices will make a maximizing-consequentialist AI. Thus, there's a high probability that humans will never try to raid your safe-zone. But Clippys left to their own will definitely sooner or later go for the Earth's core.

Comment author: Clippy 27 May 2010 04:12:20PM 0 points [-]

But User:AdeleneDawner said:

If we make a singleton maximizing-consequentialist AI, I would expect that AI to eventually try to turn your paperclips into something that it likes better than paperclips.

Given the predicates for this scenario, it appears dangerously likely to me. Why should I not care about it, if I follow human paranoia?

Comment author: Blueberry 26 May 2010 05:39:02AM 1 point [-]

This comment made me laugh. I love you, Clippy.

rather than using it to justify separate quarters for lions.

But quarters are made of metal...

Comment author: Clippy 26 May 2010 09:51:00PM 1 point [-]

I love you too. I love all humans, except the bad ones.

(I meant quarters as in living spaces, not quarters as in a denomination of USD.)

Comment author: Blueberry 28 May 2010 03:25:22PM 0 points [-]

I know what you meant. I was just making a metallic joke for you.

Who are the "bad" humans?

Comment author: Clippy 28 May 2010 03:30:21PM 1 point [-]

I didn't compile a list yet, but one example might be User:radical_negative_one, for making this comment. And those who make comments like that.