A lot of rationalist thinking about ethics and economy assumes we have very well defined utility functions - knowing exactly our preferences between states and events, not only being able to compare them (I prefer X to Y), but assigning precise numbers to every combinations of them (p% chance of X equals q% chance of Y). Because everyone wants more money, you should theoretically even be able to assign exact numerical values to positive outcomes in your life.
I did a small experiment of making a list of things I wanted, and giving them point value. I must say this experiment ended up in a failure - thinking "If I had X, would I take Y instead", and "If I had Y, would I take X instead" very often resulted in a pair of "No"s. Even thinking about multiple Xs/Ys for one Y/X usually led me to deciding they're really incomparable. Outcomes related to similar subject were relatively comparable, those in different areas in life were usually not.
I finally decided on some vague numbers and evaluated the results two months later. My success on some fields was really big, on other fields not at all, and the only thing that was clear was that numbers I assigned were completely wrong.
This leads me to two possible conclusions:
- I don't know how to draw utility functions, but they are a good model of my preferences, and I could learn how to do it.
- Utility functions are really bad match for human preferences, and one of the major premises we accept is wrong.
Anybody else tried assigning numeric values to different outcomes outside very narrow subject matter? Have you succeeded and want to share some pointers? Or failed and want to share some thought on that?
I understand that details of many utility functions will be highly personal, but if you can share your successful ones, that would be great.
It's extensively discussed and agreed upon, that that is how we (for certain definitions of "we") would like it to be, and it certainly has desirable properties for say, building Friendly AI, or any AI that doesn't wirehead. And it is certainly a property of the human brain that it orients its preferences towards what it believes is the outside world - again, it has good consequences for preventing wireheading.
But that doesn't make it actually true, just useful.
It's also pretty well established as a tenet of e.g., General Semantics, that the "outside world" is unknowable, since all we can ever consciously perceive is our map. The whole point of discussing biases is that our maps are systematically biased -- and this includes our preferences, which are being applied to our biased views of the world, rather than the actual world.
I am being descriptive here, not prescriptive. When we say we prefer a certain set of things to actually be true, we can only mean that we want the world to not dispute a certain map, because otherwise we are making the supernaturalist error of assuming that a thing could be true independent of the components that make it so.
To put it another way, if I say, "I prefer that the wings of this plane not fall off", I am speaking about the map, since "wings" do not exist in the territory.
IOW, our statements about reality are about the intersection of some portion of "observable" reality and our particular mapping (division and labeling) of it. And it cannot be otherwise, since to even talk about it, we have to carve up and label the "reality" we are discussing.
It's funny that you talk of wordplay a few comments back, as it seems that you're the one making a technically-correct-but-not-practically-meaningful argument here.
If I may attempt to explore your position: Suppose someone claims a preference for "blue ski... (read more)