You're looking at Less Wrong's discussion board. This includes all posts, including those that haven't been promoted to the front page yet. For more information, see About Less Wrong.

Dolores1984 comments on Brief Question about FAI approaches - Less Wrong Discussion

3 Post author: Dolores1984 19 September 2012 06:05AM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (42)

You are viewing a single comment's thread. Show more comments above.

Comment author: Dolores1984 19 September 2012 11:33:56PM -1 points [-]

Like I said, that part is tricky to formalize. But, ultimately, it's an individual choice on the part of the model (and, indirectly, the agent being modeled). I can't formalize what counts as a valid continuation today, let alone in all future societies. So, leave it up to the agents in question.

As for the racism thing: yeah, so? You would rather we encode our own morality into our machine, so that it will ignore aspects of people's personality we don't like? I suppose you could insist that the models behave as though they had access to the entire factual database of the AI (so, at least, they couldn't be racist simply out of factual inaccuracy), but that might be tricky to implement.

Comment author: Nisan 20 September 2012 08:09:08AM 0 points [-]

As for the racism thing: yeah, so?

Which scenario are you affirming? I'm trying to understand your intention here. Would a racist get to veto a nonracist future version of themself?

Comment author: Nisan 20 September 2012 08:07:38AM 0 points [-]

I can't formalize what counts as a valid continuation today, let alone in all future societies. So, leave it up to the agents in question.

I think you use the words "valid continuation" to refer to a confused concept. That's why it seems hard to formalize. There is no English sentence that successfully refers to the concept of valid continuation, because it is a confused concept.

If you propose to literally ask models "is this a valid continuation of you?" and simulate them sitting in a room with the future model, then you've got to think about how the models will react to those almost-meaningless words. You might as well ask them "is this a wakalix?".