byrnema comments on Attention Less Wrong: We need an FAQ - Less Wrong
You are viewing a comment permalink. View the original post to see all comments and the full post content.
You are viewing a comment permalink. View the original post to see all comments and the full post content.
Comments (108)
I'm confused by the terminology, but I think I would be a relativist objectivist.
I certainly think that morality is relative -- what is moral is agent-dependent -- but whether or not the agent is behaving morally is an objective fact about that agent's behavior, because the behavior either does or doesn't conform with that agent's morality.
But I don't think the distinction between a relativist objectivist and a relativist subjectivist is terribly exciting: it just depends on whether you consider an agent 'moral' if it conforms to its morality (relativist objectivist) or yours (relativist subjectivist).
But maybe I've got it wrong, because this view seems so reasonable, whereas you've indicated that it's rare.
The key phrase for subjectivism is "mind dependent" so if you think other people's morality comes from their minds then you are a relativist subjectivist.
I just realized I don't think people should conform to their own morality, I think people should conform to my morality which I guess would make me a subjective non-relativist.
So you believe that the word morality is a two-place word and means what an agent would want to do under certain circumstances? What word do you use to means what actually ought to be done? The particular thing that you and, to a large degree all humans would want to do under specified circumstances? Or do you believe there isn't anything that should be done other than what whatever agents exist want? Please note that that position is also a statement about what the universe ought to look like.
Yes, morality is a two-place word -- the evaluation function of whether an action is moral has two inputs: agent, action. "Agent" can be replaced by anything that conceivably has agency, so morality can be considered system-dependent, where systems include social groups and all humanity, etc.
I wouldn't say morality is what the agent wants to do, but is what the agent ought to do, given its preferences. So I think I am still using it in the usual sense.
I can talk about what I ought to do, but it seems to me I can't talk about what another agent ought to do outside their system of preferences. If I had their preferences, I ought to do what they ought to do. If they had my preferences, they ought to do what I ought to do. But to consider what they ought to do, with some mixture of preferences, isn't incoherent.
I can have a preference for what another agent does, of course, but this is different than asserting a morality. For example, if they don't do what I think is moral, I'm not morally culpable. I don't have their agency.
As far as I can tell, we don't disagree on any matter of fact. I agree that we can only optimize our own actions. I agree that other agents won't necessarily find our moral arguments persuasive. I just don't agree that the words moral and ought should be used the way you do.
To the greater LW community: Is there some way we can come up with standard terminology for this sort of thing? I myself have moved toward using the terminology used by Eliezer, but not everyone has. Are there severe objections to his terminology and if so, are there any other terminologies you think we should adopt as standard?
You're thinking of the wrong sense of objective. An objective morality, according to this article, is a morality that doesn't depend on the subject's mind. It depends on something else. I.e., if we were trying to determine what should_byrnema is, we wouldn't look at you're preferences, instead we would look somewhere else. So for example:
A nonrelativist objectivist would say that we would look at the one true universially compelling morality that's written into the fabric of reality (or something like that). So should_byrnema is just should, period.
A relativist objectivist might say (this is just one example - cultural relativism), that we would look for should_byrnema in the culture that you are currently embedded in. So should_byrnema is should_culture.
I'm not sure that subjective nonrelativism is a possibility though.
I think "subjective" means based on opinion (a mind's assessment).
If Megan-is-moral if she thinks she's moral, then the morality of Megan is subjective and depends on her mind. If Megan is moral if I think she's moral, then it's subjective and depends on my mind.
I think that whether an agent is moral or not is a fact, and doesn't depend upon the opinion/assessment of any mind. But we would still look at the agent's preferences to determine the fact. I thought this was already described by the word 'relative'.
"Subjective" has many meanings. The article uses "subjective" to mean dependent on the mind in any way. Not just a mind's assessment.
Given this definition of subjective, the article would classify your last paragraph as an example of subjective relativism.
I see. Just to clarify fully: in my last paragraph, morality depends on the mind because a mind is required for preferences and agency? Are there any exceptions to this?
yep
I dunno, my concept of mind is too fuzzy to have an answer for that.
Thanks, I do understand the framework you're using, and can now say I don't agree with it.
First, one wouldn't say that morality is subjective just because the morality of an entity depends upon its preferences and agency. Even an objective morality would usually apply moral judgments only to entities with preferences and agency.
Second, subjective should mean that Megan's action could considered moral by Fred but not moral by Tom. In other words, the morality is determined by and depends upon someone's mind. In the relative objective morality I've been speaking of, neither Megan, Fred nor Tom gets to decide if Megan's action is moral. The morality of the action is a fact of and determined by the system of Megan, her action, and the context of that action. The morality of her action is something that could be computed by something without a mind, and the morality of her action doesn't depend on the computation actually being done.
I'm not using any framework here, just definitions. The article defined relative and subjective in certain ways in order to classify moral systems, and I've just been relating how the article defines these terms. There's only semantics here, no actual inference.
Using your framing regarding what it is that we are discussing (framings cannot be avoided), perhaps I disagree with your interpretation of the phrase 'mind dependent'.
The article writes:
The article does not actually define mind-dependent. I think that by "mind-dependent", the article means that it a mind that is doing the calculation and that assigns the morality, whereas if I am understanding your position (for example), you seem to think that "mind-dependent" means that an entity being labeled moral must have a mind. In the first paragraph of my last comment, I argued that this sense of mind-dependent would make "objective morality" more or less moot, because we hardly every talk about the morality of mindless entities.
Tyrell McAllister writes:
His understanding of subjectivist also seems to interpret 'mind-dependent' as requiring a mind to do the measuring.
We seem to be talking past each other, but I'm not entirely sure where the misunderstanding is, so I'll just lay out my view of what the article says again in different terms.
A morality is subjective iff you have to look at the mind of an agent in order to determine whether they are moral. e.g., morality as preferences. A morality is objective iff you don't look at the mind of an agent in order to determine whether they are moral. For example, a single morality "written into the fabric of the universe," or a morality that says what is moral for an agent depends on where in the universe the agent happens to be (note that the former is not relative and the latter is, but I don't think we're disagreeing on what that means).
In both cases, the only type of thing being called moral is something with a mind (whatever "mind" means here). The difference is whether or not you have to look inside the mind to determine the morality of the agent.
So I'm not saying that mind dependent vs. indenpendent is the difference between having a mind and not having a mind, its the difference between looking at the mind that the agent is assumed to have and not looking at it.
Surely it's a logical possibility. Stipulate: "What's right is either X or Y, where we ask each person in the universe to think of a random integer, sum them, and pull off the last bit, 0 meaning X is right and 1 meaning Y is right."
ETA: CEV, perhaps?
Wouldn't "Everyone should do what my moral code says they should" be subjective nonrelativism? Surely there are lots of people who believe that.
I don't think the people who believe that, think that their own mental states are what determine the truth of their moral code.
Is CEV even an ethical theory? I thought it was more of an algorithm for extracting human preferences to put them in an AI.
Surely it's a de facto ethical theory, since it determines entirely what the FAI should do. But then, the FAI is not supposed to be a person, so that might make a difference for our use of 'ethical'.
hmm. Then wouldn't it be premised on subjective relativism? (relative to humans)
Yes, I'd considered that when I wrote it, but it's an odd use of 'relative' when it might be equivalent to 'the same for everyone'.
not all possible minds, just human minds
EDIT: but if you thought all possible minds had the same preferences, then it would be subjective nonrelative, wouldn't it?
Maybe, though in that unlikely event I would suspect that there's some universal law behind that odd fact about preferences, in which case I'd think it would be objective.
Well I'm not sure we need to consider merely logically possible minds, and it's logically possible that non-human minds are physically impossible.
Only in the sense that it logically possible that travel to Mars is physically impossible. The wording is deceptive.
well then, I'm just not imaginative enough!
Once you've had to argue about ethics with logicians, it becomes natural. "But what if... (completely implausible hypothesis that no one believes)" comes up a lot.