wheninrome15 comments on Open Thread: April 2010 - Less Wrong

4 Post author: Unnamed 01 April 2010 03:21PM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (524)

You are viewing a single comment's thread.

Comment author: wheninrome15 02 April 2010 12:53:16AM 1 point [-]

Is there any chance that we (a) CAN'T restrict AI to be friendly per se, but (b) (conditional on this impossibility) CAN restrict it to keep it from blowing up in our faces? If friendly AI is in fact not possible, then first generation AI may recognize this fact and not want to build a successor that would destroy the first generation AI in an act of unfriendliness.

It seems to me like the worst case would be that Friendly AI is in fact possible...but that we aren't the first to discover it. In which case AI would happily perpetuate itself. But what are the best and worst case scenarios conditioning on Friendly AI being IMpossible?

Has this been addressed before? As a disclaimer, I haven't thought much about this and I suspect that I'm dressing up the problem in a way that sounds different to me only because I don't fully understand the implications.

Comment author: PhilGoetz 02 April 2010 02:14:20AM 1 point [-]

Is there any chance that we (a) CAN'T restrict AI to be friendly per se, but (b) (conditional on this impossibility) CAN restrict it to keep it from blowing up in our faces?

First, define "friendly" in enough detail that I know that it's different from "will not blow up in our faces".

Comment author: RobinZ 02 April 2010 02:27:34AM 0 points [-]

Ooh, good catch! wheninrome15 may need to define "will not blow up in our faces" in more detail as well.

Comment author: RobinZ 02 April 2010 01:03:28AM 1 point [-]

Such an eventuality would seem to require that (a) human beings are not computable or (b) human beings are not Friendly.

In the latter case, if nothing else, there is [individual]-Friendliness to consider.

Comment author: Kevin 02 April 2010 01:16:51AM *  2 points [-]

I think human history has demonstrated that (b) is certainly true... sometimes I am surprised we are still here.

Comment author: RobinZ 02 April 2010 01:58:12AM 2 points [-]

The argument from (b)* is one of the stronger ones I've heard against FAI.

* Not to be confused with the argument from /b/.

Comment author: ata 02 April 2010 10:59:56AM 1 point [-]

Incidentally, /b/ might be good evidence for (b). It's a rather unsettling demonstration of what people do when anonymity has removed most of the incentive for signaling.

Comment author: taw 02 April 2010 01:23:24PM 2 points [-]

I find chans' lack of signaling highly intellectually refreshing. /b/ is not typical - due to ridiculously high traffic only meme-infested threads that you can reply to in 5 seconds survive. Normal boards have far better discussion quality.