Wei_Dai comments on Nonparametric Ethics - Less Wrong

27 Post author: Eliezer_Yudkowsky 20 June 2009 11:31AM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (56)

You are viewing a single comment's thread. Show more comments above.

Comment author: Wei_Dai 21 June 2009 09:12:50PM *  0 points [-]

But every extrapolation process starts with gathering detailed data points, so it confused me that you focused on "nonparametric" as a response to Robin's argument. If Robin is right, an FAI should discard most of the detailed picture of human psychology it captures during its extrapolation process as errors and end up with a few simple moral principles on its own.

Can you clarify which of the following positions you agree with?

  1. An FAI will end up with a few simple moral principles on its own.
  2. We might as well do the extrapolation ourselves and program the results into the FAI.
  3. Robin's argument is wrong or doesn't apply to the kind of moral extrapolation an FAI would do. It will end up with a transhuman morality that's no less complex than human morality.

(Presumably you don't agree with 2. I put it in just for completeness.)

Comment author: Eliezer_Yudkowsky 21 June 2009 11:08:58PM 2 points [-]

2, certainly disagree. 1 vs. 3, don't know in advance. But an FAI should not discard its detailed psychology as "error"; an AI is not subject to most of the "error" that we are talking about here. It could, however, discard various conclusions as specifically erroneous after having actually judged the errors, which is not at all the sort of correction represented by using simple models or smoothed estimators.