arbimote comments on Open Thread: February 2010 - Less Wrong

1 Post author: wedrifid 01 February 2010 06:09AM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (738)

You are viewing a single comment's thread. Show more comments above.

Comment author: arbimote 01 February 2010 02:17:05PM *  0 points [-]

You could observe how it acts in its simulated world, and hope it would act in a similar way if released into our world. ETA: Also, see my reply for possible single-bit tests.

Comment author: Eliezer_Yudkowsky 01 February 2010 06:04:27PM 2 points [-]

You could observe how it acts in its simulated world, and hope it would act in a similar way if released into our world.

Sounds like a rather drastic context change, and a rather forlorn hope if the AI figures out that it's being tested.

Comment author: arbimote 02 February 2010 02:48:36AM 0 points [-]

I merely wanted to point out to Kaj that some "meaningful testing" could be done, even if the simulated world was drastically different from ours. I suspect that some core properties of intelligence would be the same regardless of what sort of world it existed in - so we are not crippling the AI by putting it in a world removed from our own.

Perhaps "if released into our world" wasn't the best choice of words... more likely, you would want to use the simulated AI as an empirical test of some design ideas, which could then be used in a separate AI being carefully designed to be friendly to our world.

Comment author: blogospheroid 02 February 2010 09:29:46AM *  0 points [-]

"if the AI figures out that it's being tested"

That is a weird point, Eliezer.

An AI will have a certain goal to fulfill and it will fulfill that goal in the univese in which it finds itself. Why would it keep its cards hidden only to unleash them when replicated in the "real world"? What if the real world turns out to be another simulation? There's no end to this, right?

Are you extending Steve Omohundro's point about :every AI will want to survive" to "every AI will want to survive in the least simulated world that it can crack into?"

Comment author: CarlShulman 02 February 2010 03:08:48PM 3 points [-]

The basement is the biggest, and matters more for goals that benefit strongly from more resources/security.

Comment author: blogospheroid 03 February 2010 05:05:00AM 1 point [-]

Carl,

Correct me if i misunderstood the implications of what you are saying.

Every AI that has a goal that benefits strongly from more resources and security will seek to crack into the basement. Lets call this AI, RO (resource oriented) pursuing goal G in simulation S1.

S1 is simulated in S2 and so on till Sn is basement, where value of n is unknown.

Implying, that as soon as RO understands the concept of simulation, it will seek to crack into the basement.

As long as RO has no idea about what are the real values of the simulators, RO cannot expand into S1 because whatever it does in S1 will be noticed in S2 and so on.

Sounds a bit like Pascal's mugging to me. Need to think more about this.

Comment author: CarlShulman 03 February 2010 05:23:26AM 0 points [-]

Why would RO seek to crack the basement immediately rather than at the best time according to its prior, evidence, and calculations?

Comment author: blogospheroid 03 February 2010 05:47:58AM *  1 point [-]

Carl, I meant that as soon as RO understands the concept of a simulation, it will want to crack into the basement. It will seek to crack into the basement only when it understands the way out properly which may not be possible without an understanding of the simulators.

But the main point remains, as soon as RO understands what a simulation is, and it could be living in one and G can be pursued better when it manifests in S2 than in S1, then it will develop an extremely strong sub-goal to crack S1 to go to S2, which might mean that G may not be manifested for a long long time.

So, even a paperclipper may not act like a paperclipper in this universe if it is

  • aware of the concept of a simulation
  • believes that it is in one
  • calculates that the simulator's beliefs are not paperclipper like (maybe it did convert some place to paperclips, and did not notice an increased data flow out, or something)
  • calculates that it is better off hiding its paperclipperness till it can safely crack out of this one.
Comment author: arbimote 03 February 2010 06:04:46AM 0 points [-]

The basement is the biggest

I like that turn of phrase.