Incorrect comments on Selfish reasons for FAI - Less Wrong

0 Post author: snarles 17 December 2011 10:58PM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (12)

You are viewing a single comment's thread.

Comment author: Incorrect 17 December 2011 11:07:41PM 1 point [-]

the easier it is to develop FAI in comparison to unfriendly AI

What? What are your definitions of FAI and unfriendly AI?

Comment author: James_Blair 17 December 2011 11:37:39PM *  4 points [-]

Rather than unfriendly AI, I think he means a Friendly AI that's only Friendly to one person (or very few people). If we're going to be talking about this concept then we need a better term for it. My inner nerd prefers Suzumiya AI.

Comment author: antigonus 18 December 2011 12:35:58AM 4 points [-]

Genie AI?

Comment author: snarles 18 December 2011 12:31:10PM *  0 points [-]

Conceptually there is very little difference between AGI which understands the values of one human and AGI which understands a hypothetical aggregate of human values. Therefore I use FAI to refer to both concepts.

Comment author: DanArmak 18 December 2011 02:17:42PM 2 points [-]

Conceptually there is very little difference between AGI which understands the values of one human and AGI which understands a hypothetical aggregate of human values.

The big difference is in how the AI builds that aggregate from individual values. There are surely many ways to do so, and people will disagree about them (and some ways will not be wanted by anyone at all but we could get them by mistake for a non-Friendly AI). On the other hand, existing suggestions like CEV are greatly underspecified and haven't even been strictly proven to have real (or unique) solutions.

Comment author: timtyler 18 December 2011 01:59:08PM 1 point [-]

That apparently conflicts with how it is defined in Creating Friendly AI.

Comment author: snarles 18 December 2011 02:12:24PM 0 points [-]

I disagree. Why couldn't outlined procedure for creating friendly AI (3.4.4 of link) to be based on a one individual, a superposition of a small group of individuals instead of a superposition of all of humanity?

Comment author: timtyler 18 December 2011 03:56:04PM *  0 points [-]

The preference of particular individual humans may involve harming other humans.

Defining such preferences as "friendly" violates the concept as it was originallly intended, IMO.

Comment author: snarles 18 December 2011 12:28:49PM *  0 points [-]

What? What are your definitions of FAI and unfriendly AI?

Artificial general intelligence (AGI) = any intelligence with near-human capabilities of learning and innovation

Friendly AI (FAI)= an AGI which understands human values

unfriendly AI = an AGI which is not FAI

FAI will be harder to develop than unfriendly AI, but the question is how much harder?

Reasons why FAI is harder than AGI have been extensively discussed by Yudkowski and others. These include:

  • Human values are hard to understand

  • The growth mechanism of FAI must have low variability to ensure that the FAI remains friendly. But high-variability growth mechanisms (such as mutation and selection) could be the easiest path to AGI.

Comment author: Incorrect 18 December 2011 04:41:22PM 0 points [-]

FAI will be harder to develop than unfriendly AI

Doesn't this conflict with your other statement?

the easier it is to develop FAI in comparison to unfriendly AI

Comment author: snarles 18 December 2011 10:30:41PM 0 points [-]

I see your point. I have modified my comment to

the less difficult it is to develop FAI in comparison to unfriendly AI