drnickbone comments on SUDT: A toy decision theory for updateless anthropics - Less Wrong

15 Post author: Benja 23 February 2014 11:50PM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (12)

You are viewing a single comment's thread. Show more comments above.

Comment author: drnickbone 14 March 2014 07:29:20AM 1 point [-]

It strikes me that a persistently selfish agent may be somewhat altruistic towards its future selves. The agent might want its future versions to be free to follow their own selfish preferences, rather than binding them to its current selfish preferences.

Another alternative is that the agent is not only selfish but lazy... it could self-modify to bind its future selves, but that takes effort, and it can't be bothered.

Either way, it's going to take a weird sort of utility function to reproduce human selfishness in an AI.

Comment author: cousin_it 14 March 2014 10:12:56AM *  1 point [-]

Now that I think of it, caring about making more copies of yourself might be more fundamental than caring about object-level things in the world... I wonder what kind of math could be used to model this.