Stuart_Armstrong comments on Proper value learning through indifference - Less Wrong

16 Post author: Stuart_Armstrong 19 June 2014 09:39AM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (50)

You are viewing a single comment's thread. Show more comments above.

Comment author: Stuart_Armstrong 23 June 2014 01:31:57PM 2 points [-]

Indeed! An ideal moral reasoner could not predict the changes to their moral system.

I couldn't guarantee that, but instead I got a weaker condition: an agent that didn't care about the changes to their moral system.

Comment author: Psy-Kosh 23 June 2014 06:42:35PM 1 point [-]

Ah, alright.

Actually, come to think about it, even specifying the desired behavior would be tricky. Like if the agent assigned a probability of 1/2 to the proposition that tomorrow they'd transition from v to w, or some other form of mixed hypothesis re possible future transitions, what rules should an ideal moral-learning reasoner follow today?

I'm not even sure what it should be doing. mix over normalized versions of v and w? what if at least one is unbounded? Yeah, on reflection, I'm not sure what the Right Way for a "conserves expected moral evidence" agent is. There're some special cases that seem to be well specified, but I'm not sure how I'd want it to behave in the general case.