nyan_sandwich comments on Pinpointing Utility - Less Wrong

57 [deleted] 01 February 2013 03:58AM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (154)

You are viewing a single comment's thread. Show more comments above.

Comment author: [deleted] 04 February 2013 08:40:19PM 0 points [-]

That's not the issue we ran into.

Your (partial) utility functions do not contain enough information to resolve uncertainty between them. As far as I can tell, utility functions can't contain meta-preferences.

You can't just pull a correct utility function out of thin air, though. You got the utility function from somewhere; it is the output of a moral-philosophy process. You resolve the uncertainty with the same information-source from which you constructed the partial utility functions from in the first place.

No need to take the limit or do any extrapolation (except that stuff like that does seem to show up inside the moral-philosophy process.)

Comment author: Matt_Simpson 05 February 2013 03:49:11AM 1 point [-]

I think we're using "utility function" differently here. I take it to mean the function containing all information about your preferences, preferences about preferences, and higher level meta-preferences. I think you're using the term to refer to the function containing just object-level preference information. Is that correct?

Now that I make this distinction, I'm not sure VNM utility applies to meta-preferences.

Comment author: [deleted] 05 February 2013 04:41:44AM *  2 points [-]

Now that I make this distinction, I'm not sure VNM utility applies to meta-preferences.

It doesn't, AFAIK, which is why I said your utility function does not contain meta-preference and the whole moral dynamic. "utility function" is only a thing in VNM. Using it as a shorthand for "my whole reflective decision system" is incorrect use of the term, IMO.

I am not entirely sure that your utility function can't contain meta-preference, though. I could be convinced by some well-placed mathematics.

My current understanding is that you put the preference uncertainty into your ontology, extend your utility function to deal with those extra dimensions, and lift the actual moral updating to epistemological work over those extra ontology-variables. This still requires some level of preliminary moral philosophy to shoehorn your current incoherent godshatter-soup into that formal framework.

I'll hopefully formalize this some day soon to something coherent enough to be criticized.

Comment author: Matt_Simpson 26 February 2013 09:02:38PM 0 points [-]

I'll hopefully formalize this some day soon to something coherent enough to be criticized.

I look forward to it!