A lot of rationalist thinking about ethics and economy assumes we have very well defined utility functions - knowing exactly our preferences between states and events, not only being able to compare them (I prefer X to Y), but assigning precise numbers to every combinations of them (p% chance of X equals q% chance of Y). Because everyone wants more money, you should theoretically even be able to assign exact numerical values to positive outcomes in your life.
I did a small experiment of making a list of things I wanted, and giving them point value. I must say this experiment ended up in a failure - thinking "If I had X, would I take Y instead", and "If I had Y, would I take X instead" very often resulted in a pair of "No"s. Even thinking about multiple Xs/Ys for one Y/X usually led me to deciding they're really incomparable. Outcomes related to similar subject were relatively comparable, those in different areas in life were usually not.
I finally decided on some vague numbers and evaluated the results two months later. My success on some fields was really big, on other fields not at all, and the only thing that was clear was that numbers I assigned were completely wrong.
This leads me to two possible conclusions:
- I don't know how to draw utility functions, but they are a good model of my preferences, and I could learn how to do it.
- Utility functions are really bad match for human preferences, and one of the major premises we accept is wrong.
Anybody else tried assigning numeric values to different outcomes outside very narrow subject matter? Have you succeeded and want to share some pointers? Or failed and want to share some thought on that?
I understand that details of many utility functions will be highly personal, but if you can share your successful ones, that would be great.
In practice, your definition of what "liking and respecting me" means -- i.e., what evidence you expect to see in the world of that -- is part of the map, not the territory.
Suppose, for example, that your friends really and truly like and respect you... but they have to beat you up and call you names, for some other reason. Does that match what you actually value? It's out there in the territory, after all.
That is, is merely knowing that they "like and respect you" enough? Or is that phrase really just a shorthand in your map for a set of behaviors and non-behaviors that you actually value?
Note that if you argue that, "if they really liked and respected me, they wouldn't do that", then you are now back to talking about your map of what that phrase means, as opposed to what someone else's map is.
System 2 thinking is very tricky this way -- it's prone to manipulating symbols as if they were the things they're merely pointing at, as though the map were the territory... when the only things that exist in its perceptual sphere are the labels on the map.
Most of the time, when we think we're talking about the territory, we're talking about the shapes on the map, but words aren't even the shapes on the map!
We have only access to our current map to tell us about the territory, yes. But we have strong intuitions about how we would act if we could explicitly choose that our future map permanently diverge from our current map (which we currently see as the territory). If we (again by our current map) believe that this divergence would conform less to the territory (as opposed to a new map created by learning information), many of us would oppose that change even against pretty high stakes.
I mean, if Omega told me that I had to choose between