Q: Are all of a person's values comparable with each other? For example, is a candlelit dinner comparable to a sunset walk on a beach?
A: Of course. You can ask the person to choose between these two things. Their answer will give you information about what they value more.
Q: What if the person can't choose?
A: Then they probably value these two things about equally.
Q: Okay, I have another question. Are all abstract concepts comparable to each other by weight?
A: Come again?
Q: I mean, we can ask a person: "Is one mile heavier than one hour or vice versa?" That will give you information about that person's weight function, do they assign more weight to one mile or one hour.
A: The person can't choose, because the question is nonsense.
Q: But by your own argument above, doesn't that mean they weigh these things about equally?
A: It's different, because the question about value feels more meaningful to the person, even if they can't give an answer.
Q: But a question can feel meaningful without being about anything real. For example, questions about gods and demons feel meaningful to many people. What if questions about which thing is more valued are also like that?
A: The difference is that value doesn't only manifest in answers to questions, it also manifests in what actions people choose.
Q: Do you mean, for a specific binary choice you can imagine a person in a room faced with two buttons and so on?
A: Exactly.
Q: Very well. Imagine a person in a room faced with two buttons, saying "one mile is heavier than one hour" and "vice versa".
A: Screw you!
Tedious explanation of the joke: I've long been puzzled by the argument that we can ask people to choose between things, therefore people have preferences. Today I realized how to kick that argument down: by pointing out that you can ask people anything at all. So the mere act of asking can't be evidence that the question is meaningful. Very quickly this dialogue was born, I hope you like it.
I'm not sure I understand the claim or hypothesis behind this post. It's something about the meanings of "value" and "believe" in terms of evidence from statements or button-pushes, but I don't see how it's confusing in the first place.
In my view, people have preferences, and have beliefs about causality, which are expressed through actions that (are intended to) influence future world-states. This is VERY NOISY, because brains kind of suck, and because the complexity of the real world really is too big to fit into anyone's models. Instead of
I'd say something like "people take actions and have behaviors, and to the extent they are consistent, this implies preferences". No part of "we ask" or verbalizing those preferences is required. Preferences and values are the choices people make, not the things they say.
Ah, yeah. That’s why I’m not very hopeful about AI alignment. I don’t think anyone’s even defined the problem in a useful way.
Neither humans as a class nor most humans as individuals HAVE preferences that AI is able to fulfill, or even be compatible with as they are conceived today. We MAY have mental frameworks that let our preferences evolve to survive well in an AI-containing world.