wedrifid comments on (Almost) every moral theory can be represented by a utility function - Less Wrong
You are viewing a comment permalink. View the original post to see all comments and the full post content.
You are viewing a comment permalink. View the original post to see all comments and the full post content.
Comments (42)
This would seem to be a problem only provided both of:
Without a model of "I" that includes (milder, more subtle) versions of that kind of modification as also instances of "I" there is not a problem leaving the negative utility for "I lie" at !Time.now in place. Apart from, you know, the fact that you have an agent with an idiotic absolute deontological injunction in place. But that was the specification we were implementing.
The implementation of mechanisms for what constitutes the "self" actor in any absolute-deontological injunction is ridiculously complicated and introduces all sorts of problems and potential pitfalls. However, I don't think they are complexities and pitfalls that I am arbitrarily introducing for my convenience. They seem to be actual problems intrinsic to the practical task of implementing a deontological agent in a physical universe. In fact, they are largely problems intrinsic to actually defining a deontological rule precisely.
I would grant both the factors you mention - less than 1 confidence in the impossibility of time travel and acausal influences - as valid reasons to qualitatively accept all the problems of Time.greater_than.now infinities as problems for Time.less_than.now infinities. So I would make the same claim that they can work correctly if you either have an encompassing "infallible" or a correctly defined "I". (By 'correct' I mean "the same as what you would intend it to mean if you said the rule aloud".)
I assume we are imagining a different utility function here. I am imagining a utility function defined over universe-histories where negative infinity is returned when "I lie" is true at any point in time and normal stuff (like lots of utility for delicious cookies, volcano lairs and catgirls) for all the rest. Of all the undesirable outcomes that would come from executing such a utility function lying now to avoid future lying would not be one of them. No lying behavior will ever be returned as the expected utility maximising action of that function except in cases where (according to your model) it is logically impossible to execute a behavior that has a non-zero chance of resulting in a lie. In that case behavior is and should be totally arbitrary.