tim comments on The Evil AI Overlord List - Less Wrong

27 Post author: Stuart_Armstrong 20 November 2012 05:02PM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (80)

You are viewing a single comment's thread. Show more comments above.

Comment author: tim 21 November 2012 03:57:29AM *  1 point [-]

If the AI can't tell that it is in a simulation given that it is in a simulation, why would it be able to tell that it is in the real world given that it is in the real world? If such a proposition is effective in a simulation, humans would likely just make the same threat/test in the real world. If the AI forgoes this particular opportunity at world domination because it might be in a simulation, what would make it reason differently in the real world?

ETA: and if the AI can tell if its in a simulation vs the real world then its not really a test at all.

Comment author: Khoth 21 November 2012 11:11:49AM 2 points [-]

It could act probabilistically. If it knows humans would do a simulation test, but it can't tell whether it's in the test or the real world, it could behave with probability 50% and be evil with probability 50%, which gives it a 25% of getting to achieve its evil goals.

Comment author: DanielLC 21 November 2012 11:24:30PM *  1 point [-]

I didn't mean that it couldn't tell. I meant that it wouldn't be immediately obvious. It might be best to find a good test before taking over the world.

ETA: and if the AI can tell if its in a simulation vs the real world then its not really a test at all.

The entire point of a test is to figure things out. If you didn't need a test to figure out things you could test for, nobody would ever run them.

Comment author: Luke_A_Somers 21 November 2012 02:16:41PM 0 points [-]

It could be that in the simulation it can't prove it one way or the other because we control its information flow, but once outside the box we can't, and so it can.

Comment author: DanielLC 21 November 2012 11:27:57PM 0 points [-]

It would take more than controlling information flows. Thanks to conservation of expected evidence, if it can't find evidence that it is in a simulation, then it can't find evidence that it isn't. We might be able to modify its beliefs directly, but I doubt it. Also, if we could, we'd just convince it that it already ran the test.

Comment author: Luke_A_Somers 22 November 2012 08:42:16PM 0 points [-]

That's not what conservation of expected evidence means. If the best we can do is make things ambiguous from its point of view, that's our limit. The real world could well be a place it can very easily tell is a non-simulation.

Comment author: Giles 21 November 2012 02:07:40PM 0 points [-]

and if the AI can tell if its in a simulation vs the real world then its not really a test at all.

The AI would probably assign at least some probability to "the humans will try to test me first, but do a poor job of it so I can tell whether I'm in a sim or not"

Comment author: Strange7 14 April 2014 02:08:05AM -1 points [-]

If the AI forgoes this particular opportunity at world domination because it might be in a simulation, what would make it reason differently in the real world?

Hopefully nothing. An AI that plays nice out of the fear of God is still an AI that plays nice.