Stuart_Armstrong comments on Proper value learning through indifference - Less Wrong

16 Post author: Stuart_Armstrong 19 June 2014 09:39AM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (50)

You are viewing a single comment's thread. Show more comments above.

Comment author: Stuart_Armstrong 20 June 2014 09:38:23AM *  3 points [-]

The AI will not resist its values being changed in the particular way that is specified in to trigger a U transition. It will resist other changes of value.

Comment author: AlexMennen 20 June 2014 09:09:29PM 1 point [-]

That's true; it will resist changes to its "outer" utility function U. But it won't resist changes to its "inner" utility function v, which still leaves a lot of flexibility, even though that isn't its true utility function in the VNM sense. That restriction isn't strong enough to avoid the problem I pointed out above.

Comment author: Stuart_Armstrong 21 June 2014 05:41:23AM 2 points [-]

I will only allow v to change if that change will trigger the "U adaptation" (the adding and subtracting of constants). You have to specify what processes count as U adaptations (certain types of conversations with certain people, eg) and what doesn't.

Comment author: AlexMennen 21 June 2014 04:05:42PM 1 point [-]

Oh, I see. So the AI simply losing the memory that v was stored in and replacing it with random noise shoudn't count as something it will be indifferent about? How would you formalize this such that arbitrary changes to v don't trigger the indifference?

Comment author: Stuart_Armstrong 22 June 2014 08:47:51PM 1 point [-]

By specifying what counts as an allowed change in U, and making the agent in to a U maximiser. Then, just as standard maximises defend their utilities, it should defend U(un clubbing the update, and only that update)