Do you believe in an objective morality capable of being scientifically investigated (a la Sam Harris *or others*), or are you a moral nihilist/relativist? There seems to be some division on this point. I would have thought Less Wrong to be well in the former camp.

 

Edit: There seems to be some confusion - when I say "an objective morality capable of being scientifically investigated (a la Sam Harris *or others*)" - I do NOT mean something like a "one true, universal, metaphysical morality for all mind-designs" like the Socratic/Platonic Form of Good or any such nonsense. I just mean something in reality that's mind-independent - in the sense that it is hard-wired, e.g. by evolution, and thus independent/prior to any later knowledge or cognitive content - and thus can be investigated scientifically. It is a definite "is" from which we can make true "ought" statements relative to that "is". See drethelin's comment and my analysis of Clippy.

New to LessWrong?

New Comment
146 comments, sorted by Click to highlight new comments since: Today at 7:54 AM
Some comments are truncated due to high volume. (⌘F to expand all)Change truncation settings
[-][anonymous]12y240

There seems to be some division on this point.

I might be mistaken but I got the feeling that there's not much of a division, the picture I've got of LW on meta-ethics is something along the lines of: values exist in peoples heads, those are real, but if there were no people there wouldn't be any values. Values are to some extent universal, since most people care about similar things, this makes some values behave as if they were objective. If you want to categories - though I don't know what you would get out of it, it's a form of nihilism.

An appropriate question when discussing objective and subjective morality is:

  • What would an objective morality look like? VS a subjective one?
[-]Jack12y220

People here seem to share anti-realist sensibilities but then balk at the label and do weird things for anti-realists like treat moral judgments as beliefs, make is-ought mistakes, argue against non-consequentialism as if there were a fact of the matter, and expect morality to be describable in terms of a coherent and consistent set of rules instead of an ugly mess of evolved heuristics.

I'm not saying it can never be reasonable for an anti-realist to do any of those things, but it certainly seems like belief in subjective or non-cognitive morality hasn't filtered all the way through people's beliefs.

3TimS12y
I attribute this behavior in part to the desire to preserve the possibility of universal provably Friendly AI. I don't think a moral anti-realist is likely to think an AGI can be friendly to me and to Aristotle. It might not even be possible to be friendly to me and any other person.
[-]Jack12y200

I attribute this behavior in part to the desire to preserve the possibility of universal provably Friendly AI

Well that seems like the most dangerous instance of motivated cognition ever.

0ChristianKl12y
It seems like an issue that's important to get right. Is there a test we could run to see whether it's true?
0Eugine_Nier12y
Yes, but only once. ;)
0RobertLumley12y
Did you mean to link to this comment?
-2Eugine_Nier12y
Thanks, fixed.
2[anonymous]12y
When I explain my meta-ethical standpoint to people in general, I usually avoid using phrases or words such as "there is no objective morality" or "nihilism” because there is usually a lot of emotional baggage, often times go they go “ah so you think everything is permitted” which is not really what I’m trying to convey. In a lot of cases you are absolutely correct, but there are times when I think people on LW try answer “what do I think is right?”, this becomes a question concerning self-knowledge that is e.g. to what degree I'm I aware of what motivates me or can I formulate what I value?
2Jack12y
Terms like "moral subjectivism" are often associated with 'naive undergraduate moral relativism' and I suspect a lot of people are trying to avoid affiliating with the latter.
1mwengler12y
So you don't think everything is permitted? How do you convey thinking there is no objective truth value to any moral statement and then convey that something is forbidden?
1[anonymous]12y
Sure, I can. Doing something that is forbidden, results in harsh consequences (that other agents impose), that is the only meaningful definition I can come up with. Can you come up with any other useful definition?
0mwengler12y
I like to stick with other people's definitions and not come up with my own. Merriam-Webster for example: Thanks for being my straight man! :)
1[anonymous]12y
While reading your response the first time I got a bit annoyed frankly speaking. So I decided to answer it later when I wouldn't just scream blue! I might have misinterpreted your meaning, but it seems like you present a straw man of my argument. I was trying to make concepts like forbidden and permitted pay rent - even in a world where there is no objective morality, as well as show that our - at least my - intuition about "forbiddeness" and "permittedness" is derived form the kind of consequences that they result in. It's not like something is not permitted in a group, but do not have any bad consequences if preformed.
0mwengler12y
The largest rent I can ever imagine getting from terms which are in wide and common use is to use them to mean the same things everybody else means when using them. To me, it seems coming up with private definitions for public words decreases the value of these words. There are many words used to make moral statements. When you declare that no moral statement can be objectively true, then I don't think it makes sense to redefine all these words so they now get used in some other way. I doubt you will ever convince me to agree to the redefining of words away from their standard definitions because to me that is just a recipe for confusion. I have no idea what is "straw man" about any of my responses here.
1Viliam_Bur12y
A few examples could help me understand what you mean, because right now I don't have a clue. I guess the goal is to simplify the mess as much as possible, but not more. To find a smallest set of rules that would generate a similar result.
0komponisto12y
Well said.
0bryjnar12y
I agree. I can't figure out clearly enough exactly what Eliezer's metaethics is, but there definitely seem to be latent anti-realist sympathies floating around.
5DanArmak12y
Agreed. I just posted a more detailed description of these beliefs (which are mine) here. If anyone here believes in an objectively existing morality I am interested in dialogue. Right now it seems like a "not even wrong", muddled idea to me, but I could be wrong or thinking of a strawman.

After reading lots of debates on these topics, I'm no longer sure what the terms mean. Is a paperclip maximizer a "moral nihilist"? If yes, then so am I. Same for no.

0private_messaging12y
a paperclip maximizer is something that, as you know, requires incredible amount of work put into defining what a paperclip is (if that is even possible without fixing a model). It subsequently has an incredibly complex moral system, very stupid one, but incredibly complex nonetheless. Try something like equation-solver.
0TimS12y
I suspect the OP is asking whether you are a moral realist or anti-realist.
9cousin_it12y
Okay, is a paperclip maximizer a moral realist?
[-]Jack12y150

I see no reason to think a paperclip maximizer would need to have any particular meta-ethics. There are possible paperclip maximizers that are and one's that aren't. As rule of thumb, an agent's normative ethics, that is, what it cares about, be it human flourishing or paperclips does not logically constrain it's meta-ethical views.

2cousin_it12y
That's a nice and unexpected answer, so I'll continue asking questions I have no clue about :-) If metaethics doesn't influence paperclip maximization, then why do I need metaethics? Can we point out the precise difference between humans and paperclippers that gives humans the need for metaethics? Is it the fact that we're not logically omniscient about our own minds, or is it something deeper?
3Jack12y
Perhaps I misunderstood. There are definitely possible scenarios in which metaethics could matter to a paperclip maximizer. It's just that answering "what meta-ethics would the best paperclip maximizer have?" isn't any easier than answering "what is the ideal metaethics?". Varying an agent's goal structure doesn't change the question. That said, if you think humans are just like paperclip maximizers except they're trying to maximize something else than you're already 8/10ths of the way to moral anti-realism (Come! Take those last two steps the water is fine!). Of course it's also the case that meta-ethics probably matters more to humans than paperclip maximizers: In particular metaethics matters for humans because of individual moral uncertainty, group and individual moral change, differences in between individual moralities, and the overall complexity of our values. There are probably similar possible issues for paperclip maximizers-- like how should they resolve uncertainty over what counts as a paperclip or deal with agents that are ignorant of the ultimate value of paperclips-- and thinking about them pumps my anti-realist intuitions.
2Manfred12y
Well, there's certainly that. Also, human algorithms for decision-making can feel different from simply looking up a utility - the algorithm can be something more like a "treasure map" for locating morality, looking out at the world in a way that can feel as if morality was a light shining from outside.
0TimS12y
Consider dealings with agents that have morals that conflict with your own. Obviously, major value conflicts preclude co-existence. Let's assume it is a minor conflict - Bob believes eating cow milk and beef at the same meal is immoral. It is possible to develop instrumental or terminal values to resolve how much you tolerate Bob's different value - without reference to any meta-ethical theory. But I think that meta-ethical considerations play a large role in how tolerance of value conflict is resolved - for some people, at least.
2Vladimir_Nesov12y
Not obvious. (How does this "preclusion" work? Is it the best decision available to both agents?)
-1TimS12y
Well, if I don't include that sentence, someone nitpicks by saying: I was trying preempt by making it clear that McH gets imprisoned or killed, even by moral anti-realists (unless they are exceptionally stupid).
-3hankx778712y
I would certainly say a paperclip maximizer morality falls in the former camp (objective, able to be scientifically investigated, real), although I'm not intimately familiar with the realist/anti-realist terminology.
2TimS12y
Hank, why would Clippy believe that maximizing paperclips is based on something external to its own mind? Clippy could just as easily be programmed to desire staples, and Clippy is probably intelligent enough to know that. That said, I think Jack's general point about the relationship between ethics and meta-ethics is probably right.
-2hankx778712y
Presumably Clippy has a hard-coded utility function sitting in his source code somewhere. It's a real set of 0s and 1s sitting on a disk somewhere, and we could open the source file and investigate the code. Clippy's value system is a specific objective, pre-programmed utility function that's inherent in his design and independent/prior to any knowledge or other cognitive content Clippy eventually gains or invents. And yes, it could have been easily changed (until such a time as Clippy is all grown up and would prevent such change) to make him desire staples. But then that's a different design and we'd probably better call him Stapley at that point.
2TimS12y
Ideally, one would like objective facts to be universally compelling. If Clippy shows its source code to me, or to another AGI, neither of us would update in favor of believing that paper-clip maximizing is an appropriate terminal value.
-2hankx778712y
Ah, no, I don't mean "objective morality" in the sense of something like a "one true, universal, metaphysical morality for all mind-designs" like the Socratic/Platonic Form of Good or any such nonsense. I just mean what I said above, something in reality that's mind-independent and can be investigated scientifically - a definite "is" from which we can make true "ought" statements relative to that "is". See drethelin's comment below.
5buybuydandavis12y
Clippy's code is his mind.
4Jack12y
No. The physical instantiation of a utility function is not an argument for moral realism. On the complete contrary, defining moral actions as "whatever an agents utility function says" is straight-forward, definitional, no-bones-about-it moral subjectivism. Put it this way: the paperclip maximizer is not going to approve of your behavior.
2TimS12y
Hank, I definitely don't think there's any activity that (1) can reasonably be labeled "scientific investigation" and (2) can solve the is-ought divide.
0hankx778712y
I didn't think you would :) I'm curious about the consensus on LW, though. But incidentally, what do you think of Thou Art Godshatter?
0TimS12y
First, that essay is aimed primarily at those who think dualism is required in order to talk about morality at all - obviously that's not the discussion we are having. Second, the issue is not whether there are (1) universal (2) morally relevant (3) human preferences that (4) have been created by evolution. The answer to that question is yes (i.e. hunger, sexual desire). But that alone does not show that there is a universal way for humans to resolve moral dilemma. If we study and quantify the godshatter to the point that we can precisely describe "human nature," we aren't guaranteed in advance to know that appeals to human nature will resolve every moral dilemma. If reference to human nature doesn't, then evolutionary preference doesn't prove moral realism.
0hankx778712y
I'm not sure why "universality" is really that important here. Suppose we are just talking about one person, why can't they reduce their value judgments down to their own precisely described nature to resolve every moral dilemma they face? With a read-out of their actual terminal values defined by the godshatter, they can employ the usual consequentialist expected utility calculus to solve any question, in principle.
0TimS12y
This says it better than I could.
1hankx778712y
But that's just a confusion between two different meanings "objective vs. subjective". People apparently tend to interpret "objective" as something "universal" in the sense of like some metaphysical Form of Good, as opposed to "subjective" meaning "relative to a person". That distinction is completely stupid and wouldn't even occur to me. I'm using it in the sense of, something relative to a person but still "a fact of reality able to be investigated by science that is independent/prior to any of the mind's later acquisition of knowledge/content", versus "something that is not an independent/prior fact of reality, but rather some later invention of the mind".
3Jack12y
So lets clear something up: the two attributes objective/subjective and universal/relative are logically distinct. You can can have objective relativism ("What is moral is the law and the law varies from place to place.") and subjective universalism ("What is moral is just our opinions, but we all have the same opinions"). The attribute "objective" or "subjective" in meta-ethics refers to the status of moral judgments themselves not descriptive facts about what moral judgments people actually make or the causal/mental facts that lead people to make them. Of course it is the case that people make moral judgments, that we can observe those judgments and can learn something about the brains that make them. No one here is denying that there are objective facts about moral psychology. The entire question is about the status of the moral judgments themselves. What makes it true when I say "Murder is immoral"? If your answer references my mind your answer is subjectivist, if your answer is "nothing" than you are a non-cognitivist or an error theorist. All those camps are anti-realist camps. Objectivist answers include "the Categorical Imperative" and "immorality supervenes on human suffering".
0TimS12y
Is there accessible discussion out there of why one might expect real world correlation between objectivity and universality. I see that subjective universalism is logically coherent, but I wouldn't expect it to be true - it seems like too much of a coincidence that nothing objective requires people have the same beliefs, yet people do anyway.
0Jack12y
Lots of things can cause a convergence of belief other than objective truth, e.g. adaptive fitness. But certainly, objectivity usually implies universality.

Morality is a human behavior. It is in some ways analogous to trade or language: a structured social behavior that has developed in a way that often approximates particular mathematical patterns.

All of these can be investigated both empirically and intellectually: you can go out and record what people actually do, and draw conclusions from it; or you can reason from first principles about what sorts of patterns are mathematically possible; or both. For instance, you could investigate trade either beginning from the histories of actual markets, or from principles of microeconomics. You could investigate language beginning from linguistic corpora and historical linguistics ("What sorts of language do people actually use? How do they use it?"); or from formal language theory, parsing, generative grammar, etc. ("What sorts of language are possible?")

Some of the intellectual investigation of possible moralities we call "game theory"; others, somewhat less mathematical but more checked against moral intuition, "metaethics".

Asking whether there are universal, objective moral principles is a little like asking whether there are universal, objective p... (read more)

0mwengler12y
In seems in economics without much of a stretch you have things like: * buy low and sell high * don't produce items for sale at price x that cost you y>x to produce. * don't buy a productive asset that produces c cash per year for more than c/i where i is the risk free interest rate I'm not saying I can do this trick with morality, but economics seems to produce piles of actionable results.

In summary of my own current position (and which I keep wanting to make a fuller post thereof):

If factual reality F can represent a function F(M) -> M from moral instructions to moral instructions (e.g. given the fact that burning people hurts them, F("it's wrong to hurt people")-> "It's wrong to burn people"), then there may exist universal moral attractors for our given reality -- these would represent objective moralities that are true for a vast set of different moral starting positions. Much like you reach the Sierpinski T... (read more)

0mwengler12y
That sounds likely to me. The other things that sound possible to me are that we could ultimately determine We ought not build dangerously uncontrolled (AI, nuclear reactors, chemical reactors) or tolerate those who do. We ought not get sentiences to do things for us through promises of physical pain or other disutility (threats and coercion)j We may have a somewhat different understanding of what ought means by that point, just as we have a different sense of what time and position and particle and matter and energy mean from having done physics. For example, "ought" might be how we refer to a set of policiies that produce an optimum tradeoff between the short term interests of of the individual and the short terim interests of the group, perhaps designed so that a system in which the individuals followed these rules would have productivity growth at some maximum rate, or would converge on some optimum measure of individual freedom or both. Maybe "ought" would suggest such a wide ranging agreement that individuals who don't follow these rules must be adjusted or restrained because their cost unmodified or unrestrained to the group is so clearly "unfair," even possibly dangerous. I am not even Leonardo da Vinci here trying to describe what the future of science might look like. I am some tradesman in Florence in da Vinci's time trying to describe what the future of science might look like. My point isn't that any of the individiual details should be the ones we learn when we finally perfect "the moral method" (in analogy with the scientific method), but rather that the richness of what COULD happen makes it very hard to say never, and that someone being told about the possibilities of "objective science" 1000 years ago would have been pretty justified in saying "we will never know whether the sun will rise tomorrow, we will never be able to derive "will happen" from "did happen," (which I take to be the scientific analogy of can't derive "ought" from "is").
-1torekp12y
Well said. I'd go further: The "is-ought" dichotomy is overrated, as are the kindred splits between normative and descriptive, practice and theory, etc. I suggest that every "normative" statement contains some "descriptive" import and vice versa. For example, "grass is green" implies statements like "if you want to see green, then other things being equal you should see some grass", and "murder is immoral" implies something like "if you want to be able to justify your actions to fellow humans in open rational dialogue, you shouldn't murder." Where the corresponding motivation (e.g. "wanting to see green") is idiosyncratic and whimsical, the normative import seems trivial and we call the statement descriptive. Where it is nearly-universal and typically dear, the normative import looms large. But the evidence that there are two radically different kinds of statement - or one category of statement and a radically different category of non-statement - is lacking. When philosophers try to produce such evidence, they usually assume a strong form of moral internalism which is not itself justifiable.
[-][anonymous]12y40

I'm just wondering if you have read the metaethics sequence.

6hankx778712y
Yeah, and also the Complexity of value sequence.
[-]Jack12y30

There seems to be some confusion - when I say "an objective morality capable of being scientifically investigated (a la Sam Harris or others)" - I do NOT mean something like a "one true, universal, metaphysical morality for all mind-designs" like the Socratic/Platonic Form of Good or any such nonsense. I just mean something in reality that's mind-independent - in the sense that it is hard-wired, e.g. by evolution, and thus independent/prior to any later knowledge or cognitive content - and thus can be investigated scientifically. It is

... (read more)
4buybuydandavis12y
Me too. We're not identical, but we have similarities, and understanding them will likely allow us to better achieve mutual satisfaction. This is my disappointment with Harris. He ran a fundamentally sound project into the ditch by claiming, entirely unscientifically, that his utilitarian ideal was the one word of moral truth - Thou shalt serve the well being of all conscious creatures - that is the whole of law!
-2hankx778712y
I'm not sure what you're referring to here, but here's my comment explaining how this relates to Sam Harris. I addressed this previously, explaining that I am using 'objective' and 'subjective' in the common sense way of 'mind-independent' or 'mind-dependent' and explained in what specific way I'm doing that (that is, the proper basis of terminal values, and thus the rational basis for moral judgments, are hard-wired facts of reality that exist prior to, and independent of, the rest of our knowledge and cognition - and that the proper basis of terminal values is not something that is invented later, as a product of, and dependent on, later acquired/invented knowledge and chains of cognition). You just went on insisting that I'm using the terminology wrong purely as a matter of the meaning in technical philosophy. You do not have to demand, as you've been doing throughout this thread, that I only use words to refer to things that you want them to mean, when I am explicitly disclaiming any intimacy with the terms as they are used in technical philosophy and making a real effort to taboo my words in order to explain what I actually mean. Read the article on Better Disagreement and try to respond to what I'm actually saying instead of trying to argue over definitions. Ok, great. That's kind of what I mean, but it's more complicated than that. What I'm referring to here are actual terminal values written down in reality, which is different from 1) our knowledge of what we think our terminal values are, and 2) our instrumental values, rationally derived from (1), and 3) our faculty for moral intuition, which is not necessarily related to any of the above. To answer your previous question, One must, 1) scientifically investigate the nature of their terminal values, 2) rationally derive their instrumental values as a relation between (1) and the context of their current situation, and 3) Either arrive at a general principle or to an answer to the specific instance of m
2TimS12y
Hank, If you don't use the technical jargon, it is not clear what you mean, or if you are using the same meaning every time you use a term, or whether your meaning captures what it gestures at in a meaningful, non-contradictory way. To give a historical example, thinkers once thought they knew what infinity meant. Then different infinite sets that were "obviously" different in size were show to be the same size. But not all infinite sets were the same size. Now, we know that the former usage of infinity was confused and precise references to infinite sets need some discussion of cardinality. In short, you can't deviate from a common jargon and also complain that people are misunderstanding you - particularly when your deviations sometimes appeal to connotations of the terms that your particular usages do not justify. Edit: Remove comment re: editting
-2hankx778712y
Yes I can - if 1) I use the word in it's basic common sense way, and then, as a bonus in case people are confusing the common sense usage with some other technical meaning, 2) I specifically say "I'm not intimately familiar with the technical jargon, so here is what I mean by this", and then I explain specifically what I mean.
1TimS12y
Hank, I'm sorry - I was a little too harsh. My general difficulty is that I don't think you endorse what Jack called universal relativism. If you don't, then and don't go well together. It is the case that objective != universal, but objective things tend to cause universality. If you have a reason why universality isn't caused by objective fact in this case, you should state it.
1Jack12y
I would not equate it with anything magical or universal. Certainly people have tried to ground morality in natural facts, though it is another question whether or not any has succeeded. And certainly it is logically possible to have a morality that is objective, but relative, though few find that avenue plausible. What proponents of objective morality (save you) all agree about is that moral facts are made true by things other than people's attitudes. This does not mean that they don't think people's attitudes are relevant for moral judgments. For instance, there are moral objectivists who are preference utilitarians (probably most preference utilitarians are moral objectivists actually) they think there is one objective moral obligation: Maximize preference satisfaction (with what ever alterations they favor). In order to satisfy that obligation one then has to learn about the mental states of people. But they are moral objectivists because they think that what renders "Maximizing preference satisfaction is morally obligatory" true is that it is a brute fact about the natural world (or dictated by reason, or the Platonic Form of Good etc.). This is the path Sam Harris takes. He takes as an objective moral fact that human well-being is what we should promote and then examines how to do that through neuroscience. He isn't arguing that we should care about human well-being because there is a module in our brains for doing so. If he were his argument would be subjectivist under the usage of analytic philosophy, even though he is characterized correctly by reviewers as an objectivist. This is part of why your position was so confusing to me. You were waving the banner of someone who makes the classic objectivist mistake and calling yourself an objectivist while denying that you were making the mistake. Since you keep bringing up things like "Form of the Good" I assume you can see how hard it would be to justify objective moral judgments (in the way I'm using the ter
0torekp12y
For what it's worth, not only is your usage a common one, I think it is consistent with the way some philosophers have discussed meta-ethics. Also, I particularly like the narrow way you construct 'mind-dependent'. It seems to me that the facts that I am capable of reason, that I understand English, and that I am not blind are all "objective" in common sense speak, even though they are, in the broadest possible sense of the phrase, mind-dependent. This illustrates the need for care about what kind of mind-dependence makes for subjectivity.
-2Eugine_Nier12y
Why is this obvious? After all we do have a faculty of divination with which to perceive the Form of Truth.
[-][anonymous]12y30

There seems to be some confusion - when I say "an objective morality capable of being scientifically investigated (a la Sam Harris or others)" - I do NOT mean something like a "one true, universal, metaphysical morality for all mind-designs" like the Socratic/Platonic Form of Good or any such nonsense. I just mean something in reality that's mind-independent - in the sense that it is hard-wired, e.g. by evolution, and thus independent/prior to any later knowledge or cognitive content - and thus can be investigated scientifically

I thi... (read more)

-2hankx778712y
The latter question is the relevant one. I have many problems with his book, but I think he is fundamentally taking the perfect approach: rejecting both intrincisist religious dogmatism and subjectivist moral relativism, and putting forward a third path- an objective morality discoverable by science. You're right though, he just presupposes "well-being" as the standard and doesn't really try to demonstrate that scientifically. Eliezer's Complexity of value sequence is the only place I've seen anyone begin to approach this properly (although I have some problems with him as well).
1[anonymous]12y
I see, but as I asked before what would satisfy as "an objective morality discoverable by science."? What would the world look like if objective morality existed vs if it did not? You need to know what you are looking for, or at least have a crude sketch of how objective morality would work.
[-][anonymous]12y30

If Euthyphro's dilemma proves religious morality to be false, it also does the same to evolutionary morality: http://atheistethicist.blogspot.com/2009/02/euthyphro-and-evolutionary-ethics.html

0hankx778712y
Honestly I don't know why in the world people jump from 'an objective basis for values pre-programmed by evolution' to 'do whatever any of your intuitions say without thinking about it'. To equate those two things is completely stupid and not remotely the point here, so you're entire line of argument in this comment's thread is a wreck.
-1beriukay12y
If you're saying that we can't trust the morality that evolution instilled into us to be actually good, then I'd say you are correct. If you're saying that evolutionary ethicists believe that our brain has evolved an objective morality module, or somehow has latched onto a physics of objective morality... I would like to see examples of such arguments.
-2[anonymous]12y
I am saying evolutionary morality as a whole is an invalid concept that is irrelevant to the subject of morality. Actually, I can think of a minutely useful aspect of evolutionary morality: It tells us the evolutionary mechanism by which we got our current intuitions about morality is stupid because it is also the same mechanism that gave lions the intuition to (quoting the article I linked to) 'slaughter their step children, or to behead their mates and eat them, or to attack neighboring tribes and tear their members to bits (all of which occurs in the natural kingdom)'. If the mechanism by which we got our intuitions about morality is stupid, then we learn that our intuitions are completely irrelevant to the subject of morality. We also learn that we should not waste our time studying such a stupid mechanism.
2beriukay12y
I think I'd agree with everything you said up until the last sentence. Our brains are, after all, what we do our thinking with. So everything good and bad about them should be studied in detail. I'm sure you'd scoff if I turned your statement around on other poorly evolved human features. Like, say, there's no point in studying the stupid mechanism of the human eye, and that the eye is completely irrelevant to the subject of optics.
0[anonymous]12y
Nature exerts selective pressure against organisms that have a poor perception of their surroundings, but there is no equivalent selective pressure when it comes to morality. This is the reason why the difference between the human eye and the lion eye is not as significant as the difference between the human intuitions about morality and the lion's intuitions about morality. If evolution made the perception of the surroundings as wildly variable as that of morality across different species, I would have made an argument saying that we should not trust what we perceive and we should not bother to learn how our senses work. Similarly, if evolution had exerted selective pressure against immoral organisms, I would have agreed that we should trust our intuitions.
1mwengler12y
What an absolutely wild theory! Humans domination of the planet is totally mediated by the astonishing level of cooperation between humans. Matt Ridley in The Rational Optimist even reports evidence that the ability to trade is an evolutionary adaptation of humans that is more unique to humans even than language is. Humans are able to live together in densities without killing each other in numbers orders of magnitude higher than other primates. The evolutionarily value of an effective moral system seems overwhelmingly obvious to me, so it will be hard for me o realize where we might disagree. My claims are: it is human morality that is the basic set of rules for humans to interact. With the right morality, our interaction leads to superior cooperation, superior productivity, and superior numbers. Any one of these would be enough to give the humans with the right morality an evolutionary advantage over humans with a less effective morality. For example if we didn't have a highly developed respect for property, you couldn't hire workers to do as many things: you would spend too much protecting your property from them. If we didn't have such an orientation against doing violence against each other except under pretty limited circumstances, again, cooperative efforts would suffer a lot. It certainly seems the case that our moral intutions align much better with dogs and primates than with lions. But plenty of humans have decimated their enemies, and armies even till this day tend to rape every woman in sight.
0beriukay12y
But of course evolution made perception of the surroundings as wildly variable as morality. There are creatures with zero perception, and creatures with better vision (or heat perception, or magnetic or electric, or hearing or touch...) than we'll ever have. Even if humans were the only species with morality, arguing about variability doesn't hold much weight. How many things metabolize in arsenic? There's all kinds of singular evolutions that this argument seems to be unable to handle just because of the singularity of the case.
1Jack12y
So I certainly agree that that facts about evolution don't imply moral facts. But the way you talk seems to imply you think there are other ways to discover moral facts. But I doubt there are objective justifications for human morality that are any better than justifications for lion morality. In terms of what we actually end up valuing biological evolution (a long with cultural transmission) are hugely important. Certainly a brain module for altruism is not a confirmation of altruistic normative facts. But if we want to learn about human morality as it actually exists (say, for the purpose of programming something to act accordingly) it seems very unlikely that we would want to neglect this research area.
0Zetetic12y
I initially wrote up a bit of a rant, but I just want to ask a question for clarification: Do you think that evolutionary ethics is irrelevant because the neuroscience of ethics and neuroeconomics are much better candidates for understanding what humans value (and therefore for guiding our moral decisions)? I'm worried that you don't because the argument you supplied can be augmented to apply there as well: just replace "genes" with "brains". If your answer is a resounding 'no', I have a lengthy response. :)
0[anonymous]12y
IMO, what each of us values to themselves may be relevant to morality. What we intuitively value for others is not. I have to admit I have not read the metaethics sequences. From your tone, I feel I am making an elementary error. I am interested in hearing your response. Thanks
3Zetetic12y
I'm not sure if it's elementary, but I do have a couple of questions first. You say: This seems to suggest that you're a moral realist. Is that correct? I think that most forms of moral realism tend to stem from some variant of the mind projection fallacy; in this case, because we value something, we treat it as though it has some objective value. Similarly, because we almost universally hold something to be immoral, we hold its immorality to be objective, or mind independent, when in fact it is not. The morality or immorality of an action has less to do with the action itself than with how our brains react to hearing about or seeing the action. Taking this route, I would say that not only are our values relevant to morality, but the dynamic system comprising all of our individual value systems is an upper-bound to what can be in the extensional definition of "morality" if "morality" is to make any sense as a term. That is, if something is outside of what any of us can ascribe value to, then it is not moral subject matter, and furthermore; what we can and do ascribe value to is dictated by neurology. Not only that, but there is a well-known phenomenon that makes naive (without input from neuroscience) moral decision making: the distinction between liking and wanting. This distinction crops up in part because the way we evaluate possible alternatives is lossy - we can only use a very finite amount of computational power to try and predict the effects of a decision or obtaining a goal, and we have to use heuristics to do so. In addition, there is the fact that human valuation is multi-layered - we have at least three valuation mechanisms, and their interaction isn't yet fully understood. Also see Glimcher et al. Neuroeconomics and the Study of Valuation From that article: The mechanisms for choice valuation are complicated, and so are the constraints for human ability in decision making. In evaluating whether an action was moral, it's imperative to avoid making th
0mwengler12y
What would probably help is if you said what you thought was relevant to morality, rather than only telling us about things you think are irrelevant. It would make it easier to interpret your irrelevancies.
-1AlonzoFyfe12y
Evolutionary Biology might be good at telling us what we value. However, as GE Moore pointed out, ethics is about what we SHOULD value. What evolutionary ethics will teach us is that our mind/brains are maleable. Our values are not fixed. And the question of what we SHOULD value makes sense because our brains are malleable. Our desires - just like our beliefs - are not fixed. They are learned. So, the question arises, "Given that we can mold desires into different forms, what SHOULD we mold them into?" Besides, evolutionary ethics is incoherent. "I have evolved a disposition to harm people like you; therefore, you deserve to be harmed." How does a person deserve punishment just because somebody else evolved a disposition to punish him. Do we solve the question of gay marriage by determining whether the accusers actually have a genetic disposition to kill homosexuals? And if we discover they do, we leap to the conclusion that homosexuals DESERVE to be killed? Why evolve a disposition to punish? That makes no sense. What is this practice of praise and condemnation that is central to morality? Of deserved praise and condemnation? Does it make sense to punish somebody for having the wrong genes? What, according to evolutionary ethics, is the role of moral argument? Does genetics actually explain such things as the end of slavery, and a woman's right to vote? Those are very fast genetic changes. The reason that the Euthyphro argument works against evolutionary ethics because - regardless of what evolution can teach us about what we do value, it teaches us that our values are not fixed. Because values are not genetically determined, there is a realm in which it is sensible to ask about what we should value, which is a question that evolutionary ethics cannot answer. Praise and condemnation are central to our moral life precisely because these are the tools for shaping learned desires - resulting in an institution where the question of the difference between right
0mwengler12y
Its lunchtime so for fun I will answer some of your rhetorical questions. Unless GE Moore is either an alien or an artificial intelligence, he is telling us what we should value from a human brain that values things based on its evolution. How will he be able to make any value statement and tell you with a straight face that his valuing that thing has NOTHING to do with his evolution? My disposition to harm people is triggered approximately proportionally to my judgement that this person has or will harm me or someone I care about. My disposition doesn't speak, but neither does my disposition to presume based on experience that the sun will rise tomorrow. What does speak says about the second that being able to predict the future based on the past is an incredibly effective way to understand the universe, so much so that it seems the niverse's continuity from the past to the future is a feature of the universe, not just a feature of the tools my mind has developed to understand the universe. About my incoherent disposition to harm someone who is threatening my wife or my sister, I would invite you to consider life in a society where this disposition did not exist. Violent thieves would run roughshod over the non-violent, who would stand around naked, starving, and puzzled: "what can we do about this after all?" This sentence seems somewhat incoherent but I'll address what I think are some of the interesting issues it evokes, if not quite brings up. First, public open acceptance of homosexuality is a gigantic and modern phenomenon. If nothing else, it proves that an incredibly large number of humans DO NOT have any such genetic urge to kill homosexuals, or even to give them dirty looks when walking by them on the street for that matter. So if there is a lesson about concluding moral "oughts" from moral "is-es" here, it is that anybody who previously conclude that homicidal hatred of homosexuals was part of human genetic moral makeup was using insanely flawed meth
0Zetetic12y
First, I do have a couple of nitpicks: That depends. See here for instance. This depends on what you mean by "punish". If by "punish" you mean socially ostracize and disallow mating privileges, I can think of situations in which it could make evolutionary sense, although as we no longer live in our ancestral environment and have since developed a complex array of cultural norms, it no longer makes moral sense. In any event, what you've written is pretty much orthogonal to what I've said; I'm not defending what you're calling evolutionary ethics (nor am I aware of indicating that I hold that view, if anything I took it to be a bit of a strawman). Descriptive evolutionary ethics is potentially useful, but normative evolutionary ethics commits the naturalistic fallacy (as you've pointed out), and I think the Euthyphro argument is fairly weak in comparison to that point. The view you're attacking doesn't seem to take into account the interplay between genetic, epigenetic and cultural/mememtic factors in how moral intuitions are shaped and can be shaped. It sounds like a pretty flimsy position, and I'm a bit surprised that any ethicist actually holds it. I would be interested if you're willing to cite some people who currently hold the viewpoint you're addressing. Well, really it's more neuroscience that tells us that our values aren't fixed (along with how the valuation works). It also has the potential to tell us to what degree our values are fixed at any given stage of development, and how to take advantage of the present degree of malleability. Of course; under your usage of evolutionary ethics this is clearly the case. I'm not sure how this relates to my comment, however. I agree that it's pretty obvious that social reinforcement is important because it shapes moral behavior, but I'm not sure if you're trying to make a central point to me, or just airing your own position regardless of the content of my post.
0[anonymous]12y
Yay for my favorite ethicist signing up for LessWrong!

Reading your edit... I believe that there exists some X such that X developed through natural selection, X does not depend on any particular knowledge, X can be investigated scientifically, and for any moral intuition M possessed by a human in the real world, there's a high probability that M depends on X such that if X did not exist, M would not exist either. (Which is not to say that X is the sole cause of M, or that two intuitions M1 and M2 can't both derive from X such that M1 and M2 motivate mutually exclusive judgments in certain real-world situation... (read more)

Well, an awful lot of what we think of as morality is dictated, ultimately, by game theory. Which is pretty universal, as far as I can tell. Rational-as-in-winning agents will tend to favor tit-for-tat strategies, from which much of morality can be systematically derived.

4Lightwave12y
Not all of it, though, because you still need some "core" or "terminal" values that you use do decide what counts as a win. In fact, all the stuff that's derived from game theory seems to be what we call instrumental values, and they're in some sense the less important ones, the larger portion of the arguments about morality end up being about those "terminal" values, if they even exist.
1[anonymous]12y
You are talking about different things Dolores is talking about "why should I cooperate instead of cheating" kind of morality. You on the other hand are talking about meta-ethics, that is what is the meaning of right and wrong, what is value etc.
1Dolores198412y
Indeed. Terminal values are also... pretty personal, in my book. Very similar across the whole of neurologically intact humanity, maybe, but if someone's are different from yours, good bloody luck talking them out of them.

I believe in objective relative moralities.

0TimS12y
What does believing that entail?
7drethelin12y
I think different intelligent entities will have different values, but that it's objectively possible to determine what these are and what actions are correct for which ones. I also think most people's stated values are only an approximation of their actual values.
2TimS12y
I agree that it is possible to figure out an agent's terminal values by observing their behavior and such, but I don't understand what work the word "objectively" is doing in that sentence.
2Jack12y
Most people, as this thread has exhibited, don't understand what the word means or at least not what it means in phases like "objective moral facts".
1TimS12y
Given the amount of discussion of applied-morality concepts like Friendliness and CEV, I had higher expectations.
0drethelin12y
Basically it means that even though moralities may be subjective I think statements like "that's wrong" or "that's the right thing to do" are useful, even if at base meaningless.
2mwengler12y
The idea that a meaningless statement can be useful represents a fundamental misunderstanding of what the word "meaningless" means. If a statement is useful, it must have meaning, or else there would be nothing there to use.
0Jack12y
I think he means "don't refer to anything" rather than "meaningless".
0hankx778712y
This is exactly what I mean.
-1Jack12y
"Objective" means "mind-independent" so if you're looking at someone's mind to determine those values they're, by definition, subjective. When we use the words "objective" and "subjective" in meta-ethics we're almost always using them in this way and now questioning, say, whether or not there are objective facts about other people's minds.
2AlonzoFyfe12y
If "objective" is "mind independent", then are facts ABOUT minds not objective? We cannot have a science that discusses, for example, how the pre-frontal lobe functions because no such claim can be mind-independent? For every so-called subjective statement, there is an objective statement that says exactly the same thing from a different point of view. If I say, "spinich, yumm" there is a corresponding objective statement "Alonzo likes spinich" that says exactly the same thing. So, why not just focus on the objective equivalent of every subjective statement? Why pretend that there is a difference that makes any difference?
3Jack12y
Because it makes a huge difference in our understanding of morality. "Alonzo expresses a strong distaste for murder" is a very different fact than "Murder is immoral" (as commonly understood), no? ETA: Of course, given that I don't think facts like "murder is immoral" exist I'm all about focusing on the other kind of fact. But it's important to get concepts and categories straight because those two facts are not necessarily intensionally or extensionally equivalent.
0AlonzoFyfe12y
Yes. Water is made up of two hydrogen and an oxygern atom is a different fact than the earth and venus are nearly the same size. It does not bring science to its knees.
-1Jack12y
And the next time someone says that there are astronomical facts about the chemical make-up of water I will correct them as well. Which is to say I don't know what your point is and can only imagine you think I am arguing for something I am not. Perhaps it's worth clarifying things before we get glib?
-1AlonzoFyfe12y
In which case, you will be making a point - not that there are different facts, but that there are different languages. Of course, language is an invention - and there is no natural law that dictates the definition of the word "astronomy". It is merely a convention that we have adopted a language in which the term "astronomy" does not cover chemical facts. But we could have selected a different language - and there is no law of nature dictating that we could not. And, yet, these facts about language - these facts about the ways we define our terms - does not cause science to fall to its knees either. So, what are you talking about? Are you talking about morality, or are you talking about "morality"?
1Jack12y
I suppose that is true... but surely that doesn't render the word meaningless? In the actual world where words mean the things they mean and not other things that they could have meant in a world with different linguistic convention "astronomy" still means something like "the study of celestial bodies", right? Surely people asking for astronomical facts about airplanes, as if they were celestial bodies is a sign of confusion and ought to be gently corrected, no? Where in the world did you get the notion that I wanted science on its knees or that I thought it was? I'm as kinky as the next guy but I quite like science where it is. I'm completely bamboozled by this rhetoric. Do you take me for someone who believes God is required for morality or some other such nonsense? If so let me be clear: moral judgments are neither natural nor supernatura objectivel facts. They are the projection of an individuals preferences and emotions that people mistake for externally existing things, much as people mistake cuteness as an intrinsic property of babies when in fact it is simply the projection of our affinity for babies that makes them appear cute-to-us. That does not mean that there are not facts about moral judgments or that science is not on strong and worthy grounds when gathering such facts. My chief concern in my initial comment to which you replied was getting everyone straight on what the meta-ethical terminology means. People enjoy freelancing with the meanings of words like "objective", "subjective", and "relative" and it creates a terrible mess when talking about metaethics because no one know what anyone else is talking about. I didn't have any kind of straightforward factual disagreement with the original commenter, bracketing the fact that I was quite sure what their position was and if they in fact thought they had succeeded in solve a two-thousand old debate by discovering and objective foundation for morality when they had in fact just rediscovered moral subj
1AlonzoFyfe12y
Here is the general form of my argument. A person says, "X" is true of morality or of "moral judgments" in the public at large. This brings the talk of an objective morality to its knees. I answer that X is also true if science "or of "truth judgments" in the public at large. But it does not bring all talk of objectivity n science to its knees. Therefore, the original argument is invalid. A case in point: whether somethis is moral depends on your definition of moral. But there is no objective way to determine the correct definition of "moral". Therefore, there is no chance of an objective morality. Well, whether Pluto is a planet depends on your definition of "planet". There is no way todetermine an onjectively correct definition of "planet". Yet, planetology remains a science. Yes, many moral judgments are projections of an individual's likes and dislikestreated as intrinsic properties. But, then, many of their perceptions and observations are theory-laden. This does not eliminate the possibility of objectivity in science. We simply incorporate these facts about our perceptions into our objective account. The original post to which I responded did not provide a helpful definition. Defining "subjective" as "mind independent" denies the fact that minds are a part of the real world, and we can make objectively true and false claims about minds. Values may not exist without minds, but minds are real. They are a part of the world. And so are values. Every "subjective" claim has an "objective" claim that says exactly the same thing.
-1Jack12y
Of course I never made such an argument, so this rebuttal is rather odd. Your point of course leads to the question: what does make science objective. I would argue for two candidates though some might say they are the same an I'm happy to here others. 1) Scientific theories make predictions about our future experiences, constraining them. When a scientific theory is wrong we have unexpected experiences which lead us to reject or revise that theory. 2) Science reveals the universe's causal structure which gives us power to manipulate one variable which in turn alters another. If we are unable to do this as our theory expects we reject or revise that theory. The process leads to ever-more effective theories which, at their limit, model objective reality. This it seems to me is how science is objective, though again, I'm happy to hear other theories. Now. What are the predictions a moral theory makes? What experiments can I run to test it? Whether or not Pluto is a planet might not have an "objective definition" (what ever that means) but it sure as heck has an objective trajectory through space that can be calculated with precision using Newtonian physics. You can specify a date and an astronomer can tell you Pluto's position at that date. But there is no objective method for determining what a person should and should not do in an ethical dilemma. No, my position is that all moral judgments are these kinds of projections. If there were ones that weren't I wouldn't be an anti-realist This is neither here nor there. Scientific observations are theory-laden and to some point under-determined by the evidence. But ethical theories are in no way constrained by any evidence of any kind. You mean "objective" and no it doesn't. It just denies that moral judgments are part of the world outside the mind. It does not deny that. I agree that values exist, I just think they're subjective. You're either using these words differently than I am or you're totally wrong. I'm
0mwengler12y
I think the fact that astronomy means astronomy and not chemistry among rational conversationalists is as significant as the fact that the chess piece that looks sort of like a horse is the one rational chess players use as the knight. I don't think there is anything particularly significant in almost all labels, they're positive use is that you can manipulate concepts and report on your results to others using them.
0Jack12y
But try to move your pawn like a knight and see what happens.
0mwengler12y
Not quite, I don't think. If you are looking at different well-functioning well-informed minds to get the truth value of a statement, and you get different results from different minds, then the statement is subjective. If you can "prove" that all well-functioining well-informed minds would give you the same result, then you have "proved" that the statement is objective. In principle I could look at the mind of a good physicist to determine whether electrons repel each other, and the fact that my method for making the determination was to look at someone's mind would not be enough to change the statement "electrons repel each other" into a subjective statement.
0Jack12y
It's not about the method of discovery but truth-making features. You could look at the mind of a good physicist to determine whether electrons repel each other but that's not what makes "electrons repel each others" true. In contrast, what makes a moral judgment true according to subjectivism is the attitudes of the person who makes the moral judgment.
0Will_Newsome12y
Does an ontologically privileged transcendental God count as a mind? 'Cuz you'd think meta-ethical theism counts as belief in objective moral truths. So presumably "mind-independent" means something like "person-mind-or-finite-mind-independent"?
1Jack12y
Divine command theories of morality are often called "theological subjectivism". That's another example of a universal but subjective theory. But, say, Thomist moral theory is objectivist (assuming I understand it right).
0mwengler12y
THat's funny, the wikipedia article listed 'most religiously based moral theories' as examples of moral realism.
1Jack12y
Most religiously based moral theories aren't divine command theory, as far as I know.

Objective morality? Yes, in the sense that game theory is objective. No, in the sense that payoff matrices are subjective.

I believe it is poosible to scientifically determine whether people generally have many and strong reasons to promote or inhibit certain desires through the use of social tools such as praise, condemnation, reward, and punishment. I also believe that this investigation would make sense of a wealth of moral practices such as the three categories of action (obligation, prohibition, and non-obligatory permission), excuse, the four categories of culpability (intentional, knowing, reckless, negligent), supererogatory action, and. - of course - the role of praise, condemnation, reward, and punishment.

[-]TCB12y10

I agree with what seems to be the standard viewpoint here: the laws of morality are not written on the fabric of the universe, but human behavior does follow certain trends, and by analyzing these trends we can extract some descriptive rules that could be called morals.

I would find such an analysis interesting, because it'd provide insight into how people work. Personally, though, I'm only interested in what is, and I don't care at all about what "ought to be". In that sense, I suppose I'm a moral nihilist. The LessWrong obsession with develop... (read more)

I suspect that there exists an objective morality capable of being investigated, but not using the methods commonly known as science.

What we currently think of as objective knowledge comes from one of two methods:

1) Start with self-evident axioms and apply logical rules of inference. The knowledge obtained from this method is called "mathematics".

2) The method commonly called the "scientific method". Note that thanks to the problem of induction the knowledge obtained using this method can never satisfy method 1's criterion for knowled... (read more)

0pragmatist12y
This is pretty close to my view on the matter.

I'd be extremely surprised if there turned out to be some Platonic ideal of a moral system that we can compare against. But it seems fairly clear to me that the moral systems we adopt influence factors which can be objectively investigated, i.e. happiness in individuals (however defined) or stability in societies, and that moral systems can be productively thought of as commensurable with each other along these axes. Since some aspects of our emotional responses are almost certainly innate, it also seems clear to me that the observable qualities of moral... (read more)

[-]Sly12y10

I am an anti-realist, and pretty much find myself agreeing with DanArmak and Jack.

Until a few days ago I would have said I'm a nihilist, even though a few days ago I didn't know that was the label for someone who didn't believe that moral statements could be objective facts.

Now I would say a hearty "I don't know" and assign almost a 50:50 chance that there are objective moral "ought" statements.

Then in the last few days I was reminded that 1) scientific "objective facts" are generally dependent on unprovable assumptions, like the utiliity of induction, and the idea that what a few electrons did last thu... (read more)

0fubarobfusco12y
One might call induction an "undeniable assumption" instead: we cannot do without it; it's part of what we are. As a matter of human nature (and, indeed, the nature of all other animals (and computer programs) capable of learning), we do use induction, regardless of whether we can prove it. Some of the best evidence for induction might be rather crudely anthropic: we, who implicitly and constantly use induction, are here; creatures with anti-inductive priors are not here.

Objective morality is like real magic - people who want real magic just aren't satisfied with the magic that's real.

Moral relativism all the way. I mean something by morality, but it might not be exactly the same as what you mean.

Of course, moral relativism doesn't single out anything (like changing other people) that you shouldn't do, contrary to occasional usage - it just means you're doing so for your own reasons.

Nor does it mean that humans can't share pretty much all their algorithms for finding goals, due to a common heritage. And this would make humans capable of remarkable agreement about morality. But to call that an objective morality would be stretching it.

[-][anonymous]12y-20

The capability to be scientifically investigated is entirely divorced from the existence or discoverability of the thing scientifically investigated. We can devote billions of years and dollars in searching for things that do not and never did and never will exist. If investigation guaranteed existence, I'd investigate a winning lottery ticket in my wallet twice a day.

"Might is Right" is a morality in which there is no distinction between is and ought.

Natural selection favours some moralities more than other. The ones we see are those that thrive. Moral relativism mostly appears to ignore such effects.

-1TimS12y
It isn't an accident that there are no universal mandatory incest moralities out there in recorded history. That's not actually enough to prove moral realism is true. In short, there are universal morally relevant human preferences created by evolution (i.e. hunger, sex drive). That doesn't show that evolutionarily created preferences resolve all moral dilemmas.
0mwengler12y
The scientific method has hardly resolved all scientific dilemmas. So if there are real things in science, 'resolving all dilemmas' is not a requirement for scientific realism, so it would seem it shouldn't be a requirement for moral realism. "Descriptive" statements about morality (e.g. 'some, but not all, people think incest is wrong') is objective. The only real question is whether "normative" ethics can be objective. 'people think incest is wrong' is a descriptive statement. 'incest is wrong' is a normative statement. The moral realism question is really whether any normative statement can be objectively true. The intuition pump for thinking "maybe yes" comes not from incest statements, but rather I think from statements like "humans shouldn't pick an 8 year old at random and chop off his limbs with a chainsaw just to see what that looks like." Incest statements are like pumping your intuition about scientific realism by considering statements like "Wave Function Collapse is how we get probabilistic results in real experiments." If you are wondering whether there is ANY objective truth, start with the obvious ones like "the sun will rise tomorrow" and "hacking the arms of reasonably chosen children to see what that looks like is wrong."
0[anonymous]12y
It's hard for me to reconcile this statement with your response to timtyler above. I agree with your response to him. But consider the following assertion: "Every (moral) decision a human will face has a single choice that is most consistent with human nature." To me, that position implies that moral realism is true. If you disagree, could you explain why?
-1timtyler12y
Resolving all moral dilemmas is not really part of moral realism, though. For instance, checking with S.E.P., we have: It's easy to come up with some true claims about morality. Morality is part of biology - and is thus subject to the usual forms of scientific enquiry.
0mwengler12y
This is a bad example, because "moral realism" really refers to normative moral statements, not descriptive ones. I don't think there is any interesting controversy in describing what people think is wrong. The interesting controversy is whether anything is actually wrong or not. THe problem is with "Morality is part of biology" is it is ambiguous at best, many people would see that as a descriptive statement, not telling them that "therefore you ought to do what your biology tells you to do." Best to work with unambiguous statements since the requirement is "at least some." "Killing randomly chosen children to see what it feels like is wrong" is a normative moral statement, that if objectively true means morality realists are right. "Most people think killing randomly chosen childern to see what it feels like is wrong" is a descriptive statement that is objective, but doesn't tell you whether you ought to kill randomly chosen children or not.
0timtyler12y
I think, at this point, a scientist would ask what you actually meant by normative moral statements. I.e. what you mean by "right" and "wrong". I figure if you are sufficiently clear about that, the issue is dissolved, one way or the other.
0mwengler12y
Would that scientist be hoping I had something to add beyond what is in wikipedia? Because unless that scientist tells me what he doesn't understand about normative moral statements in philosophy that isn't easily found on the web, I wouldn't know how to improve on the wikipedia article.
0timtyler12y
The article offers a bunch of conflicting definitions - from social science, economics and elsewhere. Until there's a properly-formed question, it's hard to say very much about the answer.
0mwengler12y
OK, here you go then. http://en.wikipedia.org/wiki/Normative#Philosophy
0timtyler12y
That doesn't help too much with classifying things into categories of "right" and "wrong". Either one defines these terms as relative to some unspecified agent's preferences, or one gives them a naturalistic definition - e.g. as the preferences associated with universal instrumental values. Then there's the issue of which type of definition is more practical or useful.
0mwengler12y
My point a few comments ago was that moral realism is the theory that moral statements are real, not that statements about morality are real. Statements about unicorns are real: "unicorns are cute white horses with pointy horns that can only be seen by virgins" is a real statement about unicorns. Unicorns are NOT real. Any argument or disagreement in this chain arises from what is purely some sort of disagreement about how to use some terms. I don't mean to suggest that the content of moral realism or normative vs descriptive is right or true or real, but I do have rather a thing about using words and terms and other labels in the standard ways they have been used. For whatever reason, @timtyler considers the standard definitions of either moral relativism or normative to be nonsensical or incomplete or problematic in some serious way. Bully for him. In my opinion, it makes no sense to argue against what the standard definition of various terms are by pointing out that the concepts defined have problems. Rather than redefining words like moral realism and normative that have quite a long history of meaning what wikipedia describes pretty clearly they mean, I suggest that people who want to create better concepts than these should call them something else, and not argue that the standard definitions are not the standard definitions because they are stupid or wrong or whatever.

I do not like the word "morality". It is very ambiguously defined. The only version of morality that I even remotely agree with is consequentialism/utilitarianism. I can't find compelling arguments for why other people should think this way though, and I think morality ultimately comes down to people arguing about their different terminal values, which is always pointless.

0shminux12y
Arguing about them may not be fruitful, but exposing truly terminal values tends to be illuminating even to the person instinctively and implicitly holding certain values as terminal.
0DanArmak12y
Arguing about terminal values is pointless, of course. But morality does have a good and useful definition, although some people use the word differently, which muddles the issue. 'Morality' (when I use the word) refers to a certain human behavior (with animal analogues). Namely, the judging of (human) behaviors and actions as 'right' or 'wrong'. There are specific mechanisms for that in the human brain - for specific judgments, but more so for feeling moral 'rightness' or 'wrongness' even when the judgments are largely culturally defined. These judgments, and consequent feelings and reactions, are a human universal which strongly influences behavior and social structures. And so it is interesting and worthy of study and discussion. Furthermore, since humans are capable of modifying their behavior to a large extent after being verbally convinced of new claims, it is worthwhile to discuss moral theories and principles.
0buybuydandavis12y
You're about where I am. We have some in built valuations to behavior and reactions to those evaluations. Humans, like animals, judge behavior, with varying degrees or approval/disapproval, including rewards/punishments, Where we are likely different from animals is that we judge higher order behavior as well - not just the behavior, but the moral reaction to the behavior, then the moral reaction to the moral reaction to the behavior, etc. Of course morality is real and can studied scientifically, just like anything else about us. The first thing to notice on studying is that we don't have identical morality, just as we don't have identical genes or identical histories. Some of the recent work on morality shows it comes in certain dimensions - fairness, autonomy, purity, group loyalty, etc,, and people tend to weigh these different factors both consistently for their own judgments, and differently compared to the judgments of others. I interpret that as us having relatively consistent pattern matching algorithms that identify dimensions of moral saliency, but less consistent weighting of those different dimensions. That's the funny thing is that what is termed "objective morality" is transparently nonsense once you look scientifically at morality. We're not identical - obviously. We don't have identical moralities - obviously. Any particular statistic of all actual human moralities, for any population of humans, will just be one from an infinitely many possible statistics - obviously. The attempt to "scientifically" identify the One True Statistic, a la Harris, is nonsense on stilts.
0RobertLumley12y
That's a meta-level discussion of morality, which I agree is perfectly appropriate. But unless someone is already a utilitarian, very few, if any, arguments will make ey one.
1DanArmak12y
Why would I want to make someone a utilitarian? I'm not even one myself. I am human; I have different incompatible goals and desires which most likely don't combine into a single utility function in a way that adds up to normality.

a creator would be most concerned about the continuing existence of protoplasm over the changing situations it would incur over the millennia. For the non-thinking organisms ecological effects would predominate. Sociological precepts would influence the evolution of apparent morality then idealized by the religious and the philosophical. a scientific study of morality would involve correlations between anticipated historical outcomes and societal values.