You're looking at Less Wrong's discussion board. This includes all posts, including those that haven't been promoted to the front page yet. For more information, see About Less Wrong.

David_Gerard comments on Stupid Questions Open Thread Round 2 - Less Wrong Discussion

15 Post author: OpenThreadGuy 20 April 2012 07:38PM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (208)

You are viewing a single comment's thread. Show more comments above.

Comment author: David_Gerard 20 April 2012 09:41:30PM *  1 point [-]

I must confess I do not understand what you just said at all. Specifically:

  • the second sentence: could you please expand on that?
  • I think I get that the function does not evaluate itself at all, and if you ask it just says "it's just good 'cos it is, all right?"
  • Why is this a feature? (I suspect the password is "Löb's theorem", and only almost understand why.)
  • The last bit appears to be what I meant by "therefore it deserves to control the entire future." It strikes me as insufficient reason to conclude that this can in no way be improved, ever.

Does the sequence show a map of how to build metamorality from the ground up, much as writing the friendly AI will need to work from the ground up?

Comment author: hairyfigment 20 April 2012 10:12:47PM 0 points [-]

the second sentence: could you please expand on that?

I'll try: any claim that a fundamental/terminal moral goal 'is good' reduces to a tautology on this view, because "good" doesn't have anything to it besides these goals. The speaker's definition of goodness makes every true claim of this kind true by definition. (Though the more practical statements involve inference. I started to say it must be all logical inference, realized EY could not possibly have said that, and confirmed that in fact he did not.)

I get that the function does not evaluate itself at all,

Though technically it may see the act of caring about goodness as good. So I have to qualify what I said before that way.

Why is this a feature?

Because if the function could look at the mechanical, causal steps it takes, and declare them perfectly reliable, it would lead to a flat self-contradiction by Lob's Theorem. The other way looks like a contradiction but isn't. (We think.)

Comment author: David_Gerard 20 April 2012 10:23:15PM *  0 points [-]

Thank you, this helps a lot.

Though technically it may see the act of caring about goodness as good. So I have to qualify what I said before that way.

Ooh yeah, didn't spot that one. (As someone who spent a lot of time when younger thinking about this and trying to be a good person, I certainly should have spotted this.)