A lot of rationalist thinking about ethics and economy assumes we have very well defined utility functions - knowing exactly our preferences between states and events, not only being able to compare them (I prefer X to Y), but assigning precise numbers to every combinations of them (p% chance of X equals q% chance of Y). Because everyone wants more money, you should theoretically even be able to assign exact numerical values to positive outcomes in your life.
I did a small experiment of making a list of things I wanted, and giving them point value. I must say this experiment ended up in a failure - thinking "If I had X, would I take Y instead", and "If I had Y, would I take X instead" very often resulted in a pair of "No"s. Even thinking about multiple Xs/Ys for one Y/X usually led me to deciding they're really incomparable. Outcomes related to similar subject were relatively comparable, those in different areas in life were usually not.
I finally decided on some vague numbers and evaluated the results two months later. My success on some fields was really big, on other fields not at all, and the only thing that was clear was that numbers I assigned were completely wrong.
This leads me to two possible conclusions:
- I don't know how to draw utility functions, but they are a good model of my preferences, and I could learn how to do it.
- Utility functions are really bad match for human preferences, and one of the major premises we accept is wrong.
Anybody else tried assigning numeric values to different outcomes outside very narrow subject matter? Have you succeeded and want to share some pointers? Or failed and want to share some thought on that?
I understand that details of many utility functions will be highly personal, but if you can share your successful ones, that would be great.
It's not.
First, the only way it can be one is if "natural category" has the reductionist meaning of "a category based on distinctions that humans are biased towards using as discriminators", rather than "a category that 'naturally' exists in the territory". (Categories are abstractions, not physical entities, after all.)
And second, even if you do use the reductionist meaning of "natural category", then this does not in any way undermine the conclusion that "aboutness" is mind projection when you omit the entity mapping that aboutness from the description.
In other words, this argument appears to result in only one of two possibilities: either "aboutness" is not a natural category per the reductionist definition, and thus inherently a mind projection when the attribution source is omitted, or "aboutness" is a natural category per the reductionist definition... in which case the attribution source has to be a human brain (i.e., in another map).
Finally, if we entirely reject the reductionist definition of "natural category", then "natural category" is itself an instance of the mind projection fallacy, since the description omits any definition of for whom the category is "natural".
In short, QED: the argument is not sound. (I just didn't want to bother typing all this if you were going to retreat to a claim this was never your argument.)