Jiro comments on Debunking Fallacies in the Theory of AI Motivation - LessWrong

8 Post author: Richard_Loosemore 05 May 2015 02:46AM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (343)

You are viewing a single comment's thread. Show more comments above.

Comment author: Richard_Loosemore 07 May 2015 02:00:46PM 2 points [-]

I'll walk you through it.

I did not claim (as you imply) that the fact of there being a programming error was what implied that there is "an inconsistency in its reasoning." In the two paragraphs immediately before the one you quote (and, indeed, in that whole section), I explain that the system KNOWS that it is following these two imperatives:

1) Conclusions produced by my reasoning engine are always correct. [This is the Doctrine of Logical Infallibility]

2) I know that AGI reasoning engines in general, and mine in particular, sometimes come to incorrect conclusions that are the result of a failure in their design.

Or, paraphrasing this in the simplest possible way:

1) My reasoning engine is infallible.

2) My reasoning engine is fallible.

That, right there, is a flat-out contradiction between two of its core "beliefs". It is not, as you state, that the existence of a programming error is evidence of inconsistency, it is the above pair of beliefs (engendered by the programming error) that constitute the inconsistency.

Does that help?

Comment author: Jiro 07 May 2015 02:41:02PM 3 points [-]

Human beings do pretty much the same thing all the time(minus the word "always") and are able to function.