It seems you are not even building and fighting a strawman, you are fighting straw-windmills. You are so sure you are right about contentious topics, it's off-putting.
Here is what a reasonable take on moral relativism might look like, an example from Sean Carroll https://www.preposterousuniverse.com/podcast/2022/12/05/ama-december-2022/ :
...“Can you explain the difference between relativistic and constructivist foundations for morality?”
When you talk about moral relativism, what you’re saying is that morality is relative to the beliefs of some community or something like that, right? If a community decides that something is moral, then it’s moral. That’s what it means to be moral. And people who are outside the community have nothing to say about that, have no rights, no leverage, to critique it. Whereas constructivism is just saying that morality is constructed. In other words, it’s not out there in the world to be discovered. It’s something that human beings come together and individually and collectively construct on the basis of something.
And there’ll be a difference between human constructivists who think that different people might ultimately construct different, perfectly
Second, as Bramble (2017) points out, evolution just requires that pain isn’t desired, it doesn’t require the moral belief that the world would be better if you didn’t suffer. Given this, there is no way to debunk normative beliefs about the badness of pain.
Not really. Moral beliefs evolved as a consensus mechanism to improve fitness - if you didn't believe your suffering is morally relevant, you were less likely to convince others to help alleviate your suffering, thus reducing your fitness. All of the examples I see given for things that can't be explained with physical processes but can be explained by moral realism are just bad. I challenge you to give better ones.
If consciousness just is a physical phenomena, then it would be impossible to change conscious experiences without making a physical change. However, it seems eminently metaphysically possible that we could change consciousness but not make a physical change. Imagine a world physically identical to ours but in which one tomato that I see appears 1% redder than it does currently. If you think that world is possible, then consciousness is not purely physical.
...Note, I’m perfectly willing to grant t
Well-written, if wrong :P Thanks!
I was really hoping for a cogent argument for moral realism, and this is more a giant wall of text that is repetitive more than additive, and consists mostly of the same weak argument in multiple ways. "This sure feels wrong to me, and that's probably universal".
I haven't read your post due to its extreme length, but to say something in response to your opening – I think much content on LW addresses the question of confidence contra putative experts on a field and high confidence often seems warranted. The most notable recent case is LW being ahead of the curve on Covid, but also see our latest curated post.
The arguments about conciousness not being physical seem circular. If conciousness and experiences are physical, then you can't make an exact copy of brain without it experiencing conciousness, and you can in-principle transfer experiences between brains (worst case, using nanotech).
If consciousness has a physical effect, the zombie would have some other law of physics fill in and play the functional role of consciousness.
If zombie has additional physical law it's not physically identical. More generally, what do you even mean for something to be not physical and have casual effect on physical world? If there is casual effect, you can have equations about it.
No neuroscientific knowledge can communicate what it’s like to see red, for one who has never seen red.
You either define knowledge such as it can communicate what it’s like...
I think this post could be improved by including some quotations at the beginning that are representative of the dogmas of LessWrong. The Eliezer quotes are good, but I can't recall any explicit posts about moral anti-realism on LessWrong.
1. Putting stock in philosophers
My general impression is that you put far too much stock in what a majority of philosophers think. While lots of people thinking something is some evidence that it’s true, and lots of “experts” thinking something is even better evidence, I have yet to hear a compelling account of why I should think philosophers are experts at reaching correct philosophical conclusions in a reliable and consistent way, across different issues.
And, in any case, there are a variety of reasons why we should seriously doubt that what amount...
The salient point for LW is orthogonality thesis, not (alternatives to) moral realism. It's not really a philosophical point, as it's clearly possible in principle to build AIs that pursue arbitrary objectives (and don't care about their moral status). A question closer to practical relevance is about the character of goals of the more likely first AGIs, both for initial goals and what they settle on eventually.
the moral facts themselves are causally inert
If the moral facts are causally inert, then your belief in the existence of moral facts can't be caused by the moral facts!
Random remarks about consciousness:
This is of course true. The question for zombies isn’t just whether we could imagine them—I could imagine fermat’s last theorem being false, but it isn’t—but whether it’s metaphysically possible that they exist.
I can't see the difference. What exactly "metaphysically possible" means?
...But again, you could have some functional analogue that does the same physical thing that your consciousness does. Any physical affect that consciousness has on the world could be in theory caused by something else. If consciousness has an aff
All of the things done by consciousness would be done by other laws that are functionally identical to consciousness in this world, but that don’t contain any experiences.
Of course , physical identity is taken to include laws as well as material structure,.and p-zombie proponents consider them possible, in some appropriate sense, in our world with our laws.
This was responded to above — when we reflect on pain we conclude that it’s the type of thing that’s worth avoiding, that there should be less of.
When I reflect on my pain, I notice that it's the kind of thing I want to avoid. Other peoples pain doesn't exist phenomenally for me...I don't feel it. An ethical theory should modify my behaviour with regard to others somehow, so I don't think you get one out of subjective feelings alone.
Consider the world as it was at the time of the dinosaurs before anyone had any moral beliefs. Think about scenarios in which dinosaurs experienced immense agony, having their throats ripped out by other dinosaurs. It seems really, really obvious that that was bad.
"Bad" has a bunch of meanings, many of which are not morally relevant. A bad apple, or a bad movie are not moral wrongs.
There is a whole bunch of reasons to think that this isn't a moral wrong:-
It's not a an intentional act.
It's not breaking any rules.
It's nature's way: The Dyno that gets its throat ripped out was having a bad day, but the other one is getting to feed.
This is too long to provide a detailed response. I am not sure you interpret "moral realism" the same way. To me, it means something like this:
Imagine universe without any humans (or any other sentient beings). From my perspective, talking about "morality" in such universe simply does not make sense, this word does not apply to anything that exists there.
(As a hypothetical alternative, if morality was somehow encoded in the laws of physics or something like that, things could possibly be moral or immoral even in a completely dead universe, like maybe a pen...
I notice that such terms as "real", "objective" and their opposites are pretty bad at capturing the nuances of philosophical positions. It's one of the issues of conventional philosophy, the lexicon is flawed thus there are these endless arguments about definitions.
Classical LW framework of map-territory distinction is more helpful here. Some elements of the map can be wrong - have no referent on the territory and serve no utility. Some can directly (1 to 1) reference the elements of the territory. Some can reference the elements of more detailed maps, be...
(Disclaimer: didn't read the post, it is too long and I doubted it would engage with my views.)
I'm not sure how popular moral anti-realism actually is here. For example, Eliezer's position was technically moral realist, though his metaethics was kind of strange.
I'm not sure whether to classify myself as a moral realist or anti-realist. Regarding your litmus test "it's wrong to torture babies for fun" I find myself saying that it's true in a sense, but in a different sense than we normally use the word "true". How important this difference is depends on whe...
Lesswrongers seem to largely be in agreement in rejecting robust moral realism and accepting physicalism about consciousness. This is a shame, I think, because both of these views are incorrect. The thing I find most frustrating about this is they tend to be supremely confident on topics when a hefty percentage of philosophers disagree with them. 'Don't believe things that are widespread in your ingroup with super high confidence when a large percentage of philosophers disagree with you' seems to be a pretty good heuristic -- yet LessWrongers seem not to adhere to it, at least, in evaluating these views.
I'd like to say at the outset a few things. First, this clearly doesn't apply to all people on LessWrong. Second, I agree with LessWrongers on a huge number of things -- AI risks, for example, as well as the desirability of effective altruism. I'm broadly on board with the project of being less wrong. Thus, my criticism of LW is less of the idea behind it, and more of the particular sets of beliefs that actual LessWrongers tend to have. Third, most of this will be a crosspost of things I've written elsewhere on my blog -- I have no desire to reinvent the wheel when it comes to arguments for moral realism.
1 Moral Realism
0 An Introduction to moral realism
There are vast numbers of superficially clever arguments one can generate for crazy, skeptical conclusions; conclusions like that the external world doesn’t exist, we can’t know anything, memory isn’t reliable, and so on. These arguments, while interesting and no doubt useful if one ever comes across a real honest-to-god skeptic — a rather rare breed — don’t have much significance; skepticism exists as little more than a curiosity in the mind of the modern philosopher, something which takes real thought to refute, yet is not worth taking seriously as a serious set of views.
Yet there’s one1 form of extreme skepticism with actually existing trenchant advocates — real advocates who fill philosophy departments, rather than, like the external world or memory skeptic, merely being hypothetical advocates for the devil in philosophy papers. This skeptic is one who doubts that there are objective moral truths — moral facts made true not by the beliefs of any person.
Moral realism is the claim that there are true moral facts — ones that are not made true by anyone’s attitudes towards them. So if you think that the sentence that will follow this one is true and would be so even if no one else thought it was, you’re a moral realist. It’s typically wrong to torture infants for fun!
Now, no doubt to the moral anti-realist, my remarks sound harsh. How dare I compare them to the person who doubts anything can be really known.
— A hypothetical moral anti-realist
Well, in this article, I’ll explain why moral anti-realism is so implausible — while one always can accept the anti-realist conclusion, it’s always possible to bite the bullet on crazy conclusions. Yet moral anti-realism, much like anti-realism about the external world, is wildly implausible in what it says about the world.
We do not live in a bleak world, devoid of meaning and value. Our world is packed with value, positively buzzing with it, at least, if you know where to look, and don’t fall pray to crazy skepticism. Unfortunately the flip side of that is that the world is also packed full of disvalue — horrific, agonizing, pointless, meaningless suffering, suffering that flips the otherwise positive value of the hedonic register — that suffering must be eliminated as soon as possible. It is a moral emergency every second that it goes on.
In this article, I will defend moral realism. I will defend that it is, in fact, wrong to torture infants for fun — even if everyone disagreed. It’s no surprise that Moral realism is accepted by a majority of philosophers, though it’s certainly far from a universal view.
1 A Point About Methodology
Seeming is believing — as I hope to argue. Or, more specifically, if X seems the case to you, in general, that gives you some reason to think X is, in fact, the case. I’ve already addressed this in a previous article, so I’ll quote that.
Maybe you’re not a phenomenal conservative. Perhaps you think that in some cases, intuitions don’t serve as justification. However, we should all accept the following more modest principle.
Wise Phenomenal Conservatism: If P seems true upon careful reflection from competent observers, that gives us some prima facie reason to believe P.
This allows us to sidestep the main objections to phenomenal conservatism listed here.
Responding to the crazy appearances objection
But in the case of a person to whom a certain religion seems true, this is no doubt not after careful, prolonged rational reflection in which they consider all of the facts. If a very rational person considered all the facts and religion still seemed to have prima facie justification, it seems they would be justified in thinking religion is true. This objection is also diffused by Huemer’s responses to it.
There’s certainly much more to be said on this topic, only a minuscule portion of which I can discuss in this article. However, in philosophy, it’s pretty widely accepted that what seems to be the case probably is the case, all else equal, in at least most cases. One can accept epistemic particularism, for example, and still accept this modest requirement.
Responding to the alleged defeaters in the moral domain
Walter Sinnott-Armstrong argues that we need extra justification in some sorts of cases. If a person had a belief in a proposition purely as a result of self-interested motivated reasoning, their seeming wouldn’t be justified. Thus, he argues a constraint for accepting a belief to garner prima facie justification is the following
However, as Ballantyne and Thurrow note, this isn’t a blanket defeater for our moral beliefs; rather, this is only a defeater for the subset of our moral beliefs that are likely to be caused in some way by partial considerations.
So now the question is whether, in regards to the specific thought experiments I’ll appeal to in defending moral realism, they are plausibly caused by partiality. We’ll investigate this more in regard to the specific thought experiments that I’ll appeal to.
However, one thing is worth noting. Utilitarianism seems to have a plausible route to avoiding these objections. Utilitarianism is frequently chided for being too demanding, for being too impartial. Thus, this gives us a good reason to revise the intuitions of utilitarianism’s rivals, though not utilitarianism.
This principle is also too broad. Let’s imagine that all people had self-interested reasons to believe in core logical or mathematical facts. This wouldn’t mean we should reject the core truth of modus ponens or the core mathematical axioms. Perhaps it would undercut the intuition, but it wouldn’t be enough to totally eliminate the intuition.
This is one worry I have with Armstrong’s approach. He seems much too willing to divide intuitions into two distinct classes: justified and unjustified. However, justification comes in degrees. Declaring an intuition flat out justified or flat out unjustified seems to be a mistake — just like declaring a food hot or cold would be unwise, if one were attempting to make precise judgments about the average temperature of a room.
Armstrong’s next constraint is the following.
Several points are worth making. First, the intuitions I’m appealing to are very widespread — not many people lack the intuitions to which I’ll appeal. Perhaps some people end up reflectively rejecting those intuitions, but people tend to have the intuitions. Thus, we need not revise these intuitions in light of those who disagree. I’ll defend this more later.
Second, given that most philosophers are moral realists, it seems that most relevant domain experts find the intuitions appealing. If they didn’t, they almost surely wouldn’t be moral realists.
All of our decisions are clouded by emotion to some degree. That does not mean that we should abandon all of our judgments. Again, rather than seeing things as a yes/no question of whether or not our intuitions are justified, it makes far more sense to see justification as coming in degrees. The more emotional we are, the less we should trust our intuitions. However, we shouldn’t throw out all of our intuitions based merely on our omnipresent emotions.
With my traditional caveat about justification in coming in degrees, this seems mostly correct.
Ibid
2 Some Intuitions That Support Moral Realism
The most commonly cited objection to moral anti-realism in the literature is that it’s unintuitive. There is a vast wealth of scenarios in which anti-realism ends up being very counterintuitive. We’ll divide things up more specifically; each particular version of anti-realism has special cases in which it delivers exceptionally unintuitive results. Here are two cases
This first case is the thing that convinced me of moral realism originally. Consider the world as it was at the time of the dinosaurs before anyone had any moral beliefs. Think about scenarios in which dinosaurs experienced immense agony, having their throats ripped out by other dinosaurs. It seems really, really obvious that that was bad.
The thing that’s bad about having one’s throat ripped out has nothing to do with the opinions of moral observers. Rather, it has to do with the actual badness of having one’s throat ripped out by a T-Rex. When we think about what’s bad about pain, anti-realists get the order of explanation wrong. We think that pain is bad because it is — it’s not bad merely because we think it is.
The second broad, general case is of the following variety. Take any action — torturing infants for fun is a good example because pretty much everyone agrees that it’s the type of thing you generally shouldn’t do. It really seems like the following sentence is true
“It’s wrong to torture infants for fun, and it would be wrong to do so even if everyone thought it wasn’t wrong.”
Similarly, if there were a society that thought that they were religiously commanded to peck out the eyes of infants, they would be doing something really wrong. This would be so even if every single person in that society thought it wasn’t wrong.
Everyone could think it’s okay to torture animals in factory farms, and it would still be horrifically immoral.
This becomes especially clear when we consider moral questions that we’re not sure about. When we try to make a decision about whether abortion is wrong, or eating meat, we’re trying to discover, not invent, the answer. If the answer were just whatever we or someone else said it was — or if there was no answer — then it would make no sense to deliberate about whether or not it was wrong.
Whenever you argue about morality, it seems you are assuming that there is some right answer — and that answer isn’t made sense by anyone’s attitude towards it.
Let’s see whether these results can be debunked as a result of biasing factors.
I’m not particularly partial about whether the dinosaur’s suffering was bad. It has little emotional impact on me and I am not a dinosaur. Additionally, I’m not very partial on the question of whether torturing infants would be wrong even if everyone thought it wasn’t wrong — this will never affect me, and the moral facts themselves are causally inert. Thus, this judgment can’t be debunked by partiality considerations.
Very few people disagree, at least based on initial intuitions, with the judgments I’ve laid out. I did a small poll of people on Twitter, asking the question of whether it would be wrong to torture infants for fun, and would be so even if no one thought it was. So far, 82.6% of people have been in agreement.
There are some people who disagree. However, there is almost inevitable disagreement. If disagreement made us abandon our beliefs, we’d abandon our beliefs in political claims, because there’s way more disagreement about political claims than there is about the claim that it’s typically wrong to torture infants for fun.
Also, those who disagree tend to have views that I think are factually mistaken on independent grounds. Anti-realists seem more likely to adopt other claims that I find implausible. Additionally, they tend to make the error of not placing significant weight on moral intuitions. Thus, I think we have independent reasons to prefer the belief in realism.
It also seems like a lot of the anti-realists who don’t find the sentence “it’s typically wrong to torture infants for fun and would be so even if everyone disagreed” intuitive, tend to be confused about what moral statements mean — about what it means to say that things are wrong. I, on the other hand, like most moral realists, and indeed many anti-realists, understand what the sentence means. Thus, I have direct acquaintance to the coherence of moral sentences — I directly understand what it means to say that things are bad or wrong.
If it turned out that a lot of the skeptics of quantum mechanics just turned out to not understand the theory, that would give us good reason to discount their views. This seems to be pretty much the situation in the moral domain.
Additionally, given that most philosophers are moral realists, we have good reason to find it the more intuitively plausible view. If the consensus of people who have carefully studied an issue tends to support moral realism, this gives us good reason to think that moral realism is true. The wisdom of the crowds tends to be greater than that of any individual.
I’m really not particularly emotional about the notion that dinosaur suffering was bad. Nor do I have a particularly strong emotional reaction to some types of wrong actions, say tax fraud. If there was a type of tax fraud that decreased aggregate utility, I’d think it was wrong, even if everyone thought it wasn’t. I have no emotional attachment to that belief.
Additionally, we have good evidence from the dual process literature that careful, prolonged reflection tends to be what causes utilitarian beliefs — it’s the unreliable emotional reactions that causes our non-utilitarian beliefs. Thus, at best, this would give a reason to revise our non-utilitarian beliefs. I’ll quote an article I wrote on the subject.
Additionally, there are lots of moral results that seem to be backed by no emotional results. For example, I accept the repugnant conclusion, though I have no emotional attachment to doing so.
We have no reason to think that beliefs in the moral domain — particularly ones that reach reflective equilibrium — are particularly susceptible to illusion. This is especially true of the consequentialist ones.
This isn’t true of moral belief. The belief that dinosaur suffering was bad, even before any person had ever formed that thought, was formed through careful reflection on the nature of their suffering — it wasn’t on the basis of anything else.
What if the folk think differently
I’m supremely confident that if you asked the folk whether it would be typically wrong to torture infants for fun, even if no one thought it was, they’d tend to say yes. Additionally, it turns out that The Folk Probably do Think What you Think They Think.
Also, I trust the reflective judgment of myself and qualified philosophers significantly more than I trust the folk. Sorry folk!
Classifying anti-realists
Given that, as previously discussed, moral realism is the view that there are true moral statements, that are true independently of people’s beliefs about them, there are three ways to deny it.
Non-cognitivism — this says that moral statements are neither true nor false; they’re not in the business of being true or false. On this view, moral statements are not truth-apt, in that they can’t be true or false. There are lots of sentences that are not truth-apt — examples include “shut the door.” Shut the door isn’t true or false.
Error theory — this says that moral statements, much like statements about witches, try to state facts, but they are systematically false. For example, if a person says ‘witches can fly and cast spells’ they think they’re saying something true, but they falsely believe in a vast category of things that aren’t real, namely, witches. Thus, all positive statements about morality, much like all positive statements about witches according to error theory, turn out to be false.
Subjectivism — this says that moral statements are hinge on people’s attitudes towards them. There are different versions of subjectivism — they’re all implausible.
It turns out that each of these views has especially implausible results, ones not shared by the other three.
Non-cognitivism
Non-cognitivists think that moral statements are not truth apt. A non-cognitivist might think that saying murder is wrong really means boo! murder, or don’t murder! I’ve already explained why I think non-cognitivism is super implausible, which I’ll quote here.
Error Theory
Error theory says that all positive moral statements are false. Error theory is best described as in error theory, because of how sharply it diverges from the truth. It runs into a problem — there are obviously some true moral statements. Consider the following six examples.
What the icebox killers did was wrong.
The holocaust was immoral.
Torturing infants for fun is typically wrong.
Burning people at the stake is wrong.
It is immoral to cause innocent people to experience infinite torture.
Pleasure is better than pain.
The error theorist has to say that the meaning of those terms is exactly the same as what the realist thinks. The error theorist has to think that when people say the holocaust is bad, they’re actually making a mistake. However, this is terribly implausible. It really, really doesn’t seem like the claim ‘the holocaust is bad’ is mistaken.
Any argument for error theory will be way less intuitive than the notion that the Holocaust was, in fact, bad.
Let’s test these intuitions.
I’m not really that partial about many things I take to be bad. I think malaria is bad, despite not being personally affected by malaria. Similarly, I am in no way harmed by most of history’s evils — including hypothetical evils that have never been experienced, but that I recognized would be bad if experienced.
On top of this, this may be a reason to rethink the intuition somewhat, but it’s certainly not a reason to just throw out any intuition stemming from
Very few people disagree that the notion that it’s wrong to cause infinite torture is intuitive.
Being emotional does reduce the probative force of intuitions. However, it does not suffice to debunk an intuition — we cannot merely disregard intuitions because there’s some emotional impact. But also, I’m not particularly emotional when I consider suffering in the abstract. It still seems clearly bad.
The responses to four and five from above still apply.
Subjectivism
Subjectivism holds that moral facts depend on some people’s beliefs or desires. This could be the desires of a culture — if so, it’s called cultural relativism.
Cultural Relativism: Crazy, Illogical, and Accepted by no One Except Philosophically Illiterate Gender Studies Majors
Cultural relativism is — as the sub-header suggested — something that I find rather implausible. There are no serious philosophers that I know of who defend cultural relativism. One is a cultural relativist if they think that something is right if a society thinks that it is right.
Problem: it’s obviously false. Consider a few examples.
Imagine the Nazis convinced everyone that their holocaust was good. This would clearly not make it good.
Imagine there was a society that was in universal agreement that all babies should be tortured to death in a maximally horrible and brutal way. That wouldn’t be objectively good.
People often accept cultural relativism because they’re vaguely confused and want to be tolerant. But if cultural relativism is true, then tolerance is only good if supported by the broader culture. On cultural relativism, disagreeing with the norms of one’s broader culture is incoherent. Saying my culture is acting wrongly is just a contradiction in terms. Yet that’s clearly absurd.
This also means that if two different cultures argue about which norm is correct, they’re arguing about nothing. If norms are relative to a culture then there’s no fact of the matter about which culture is correct. But that’s absurd; the Nazis were worse than non-Nazis.
To quote my previous article on the subject
Individual Subjectivism
Individual subjectivism says that morality is determined by the attitude of the speaker. The statement murder is wrong means “I disapprove of murder.” There are, of course, more subtle versions, but this is likely the main version.
I’ve already given objections in my previous article on the subject.
Conclusion of this section
So, I think that the moral conclusions of moral anti-realism are absurd. It holds that wrongness either isn’t real or depends on our desires in some way. But that’s just wrong! It is well and truly wrong to torture infants to death, and it would be so even if no one agreed.
3 Irrational Desires
The argument I intend to lay out is relatively simple in its essence, relatively drab, and yet quite forceful.
1 If moral realism is not true, then we don’t have irrational desires
2 We do have irrational desires
Therefore, moral realism is true
Defending premise 1
Premise one seems the most controversial to laypersons, but it is premise 2 that is disputed by the philosophical anti-realists. Morality is about what we have reason to do — impartial reason, to be specific. These reasons are not dependent on our desires.
Morality thus describes what reasons we have to do things, unmoored from our desires. When one claims it’s wrong to murder, they mean that, even were one to desires murdering another, they shouldn’t do it — they have a reason not to do it, independent of desires.
Thus, the argument for premise one is as follows.
1 If there are desire independent reasons, there are impartial desire independent reasons
2 If there are impartial desire independent reasons, morality is objective
Therefore, morality is objective.
Premise 2 is true by definition. Premise 1 is trivial — impartial desire independent reasons are just identical to non-impartial desire independent reasons, but adding in a requirement of impartiality. This can be achieved by, for example, making decisions from behind the veil of ignorance — or some other similar system.
Thus, if you actual have reasons to have particular desires — to aim for particular things, then morality is objective. Let’s now investigate that assumption.
Defending Premise 2
Premise 2 states that there are, in fact, irrational desires. This premise is obvious enough.
Note here I use desire in a broad sense. By desire I do not mean what merely enjoys; that obviously can’t be irrational. My preference for chocolate ice-cream over vanilla ice cream clearly cannot be in error. Rather, I use desire in a broad sense to indicate one’s ultimate aims, in light of the things that they enjoy. I’ll use desire, broad aims, goals, and ultimate goals interchangeably.
Thus, the question is not whether one who prefers chocolate to vanilla is a fool. Instead, it’s whether someone who prefers chocolate to vanilla but gets vanilla for no reason is acting foolishly.
The anti-realist is in the difficult position of denying one of the most evident facts of the human condition — that we can be fools not merely in how we get what we want but in what we want in the first place.
Consider the following cases.
1 Future Tuesday Indifference2: A person doesn’t care what happens to them on a future Tuesday. When Tuesday rolls around, they care a great deal about what happens to them; they’re just indifferent to happenings on a future Tuesday. This person is given the following gamble — they can either get a pinprick on Monday or endure the fires of hell on Tuesday. If they endure the fires of hell on Tuesday, this will not merely affect what happens this Tuesday — every Tuesday until the sun burns out shall be accompanied by unfathomable misery — the likes of which can’t be imagined, next to which the collective misery of history’s worst atrocities is but a paltry, vanishing scintilla.
They know that when Tuesday rolls around, they will shriek till their vocal chords are destroyed, for the agony is unendurable (their vocal chords will be healed before Wednesday, so they shall only suffer on Tuesday). They shall cry out for death, yet none shall be afforded to them.
Yet they already know this. However, they simply do not care what happens to them on Tuesday. They do not dissociate from their Tuesday self — they think they’re the same person as their Tuesday self. However, they just don’t care what happens to themself on Tuesday.
Now you might be tempted to imagine that they don’t actually mind what happens on Tuesday — after all, they’re indifferent to what happens on Tuesday. This misses the case; they are only indifferent to what happens on future Tuesdays. When Tuesday rolls around, they will fiercely regret their decision. Yet after Tuesday is done, they will be glad that they made the decision — after all, they don’t care what happens on a future Tuesday. We can even stipulate that when it’s Tuesday, they’re hypnotized to believe it’s a Monday, so their suffering feels from the inside exactly and precisely as it would were it experienced on Monday.
This person with indifference to future Tuesdays is clearly making an error. This is not a minor, menial error. In fact, this is certainly the gravest error in human history — one which inflicts more misery than any other. However, the anti-realist must insist that, not only is it not the greatest error in human history, it isn’t an error at all.
After all, the person is making no factual error — they are perfectly aware that they will suffer on a future Tuesday. On the anti-realist account, where lies their error. They know they will suffer, yet they do not care — the suffering will be on a Tuesday.
Only the moral realist can account for their error — for their irrationality and great foolishness in aiming at unfathomable misery on Tuesday, rather than a pinprick on Monday. On the realist account — or at least the sensible realist account, no doubt some crazy natural law theorists would deny this — we all have reason to avoid future agony. This explains why it would be an error to subject oneself to infinite torture on a Tuesday. The fact that it’s a Tuesday gives one no reason to discount their suffering.
Now the anti-realist could try to avoid this by claiming that a decision is irrational if one will regret it. However, this runs into three problems.
First, if anti-realism is true then we have no desire independent reason to do things. It doesn’t matter if we’ll regret them. Thus, regrettably, this criteria fails. Second, by this standard both getting the pinprick on a single Monday and the hellish torture on Tuesday would be irrational, because the person who experiences them will regret each of them at various points. After all, on all days of the week except Tuesday, they’d regret making the decision to endure a Monday pinprick. Third, even if by stubbornness they never swayed in their verdict, that would in no way change whether they choose rightly.
2 Picking Grass: Suppose a person hates picking grass — they derive no enjoyment from it and it causes them a good deal of suffering. There is no upside to picking grass, they don’t find it meaningful or causing of virtue. This person simply has a desire to pick grass. Suppose on top of this that they are terribly allergic to grass — picking it causes them to develop painful ulcers that itch and hurt. However, despite this, and despite never enjoying it, they spend hours a day picking grass.
Is the miserable grass picker really making no error? Could there be a conclusion more obvious than that the person who picks grass all day is acting the fool — that their life is really worse than one whose life is brimming with meaning, happiness, and love?
3 Left Side Indifference: A person is indifferent to suffering that’s on the left side of their body. They still feel suffering on the left side of their body just as vividly and intensely as it would be on the right side of their body. Indeed, we can even imagine that they feel it a hundred times more vividly and intensely — it wouldn’t matter. However, they do not care about the left side suffering.
It induces them to cry out in pain, it is agony after all. But much like agony that one endures for a greater purpose, the agony one endures on a run say, they do not think it is actually bad. Thus, this person has a blazing iron burn the left side of their body from head to toe, inflicting profound agony. They cry out in pain as they do it. On the anti-realist account, they’re acting totally rationally. Yet that’s clearly crazy!
4 Four-Year-Old Children: Suppose that — and this is not an implausible assumption — there’s a four-year-old child who doesn’t want to go into a Doctor’s office. After all, they really don’t like shots. This child is informed of the relevant facts — if they don’t go into the Doctor’s office, they will die a horribly painful death of cancer. You clearly explain this to them so that they’re aware of all the relevant facts. However, the four-year-old still digs in their heels (I hear they tend to do that) and refuses categorically to go into the Doctor’s office.
It’s incredibly obvious that the four-year-old is being irrational. Yet they’ve been informed of the relevant facts and are acting in accordance with their desires. So on anti-realism, they’re being totally rational.
5 Cutting: Consider a person who is depressed and cuts themself. When they do it, they desire to cut themself. It’s not implausible that being informed of all the relevant facts wouldn’t make that desire go away. In this case, it still seems they’re being irrational.
6 Consistent Anorexia: A person desires to be thin even if it brings about their starvation. This brings them no joy. They starve themself to death. It really seems that they’re being irrational.
7 A person had consensual homosexual sex. They then become part of a religious cult. This religious cult doesn’t have any factual mistakes, they don’t believe in god. However, they think that homosexual sex is horrifically immoral and those who do it deserve to suffer, just as a base moral principle. On the anti-realist account, not only are they not mistaken, they would be fully rational to endure infinite suffering because they think they deserve it.
8 A person wants to commit suicide and know all the relevant facts. Their future will be very positive in terms of expected well-being. On anti-realism, it would be rational to commit suicide.
9 A person is currently enduring more suffering than anyone ever has in all of human history. However, while this person doesn’t enjoy suffering — they experience it the same way the rest of us do, they have a higher order indifference to it. While they hate their experience and cry out in agony, they don’t actually want their agony to end. They don’t care on a higher level. On this account, they have no reason to end their agony. But that’s clearly implausible.
10 A person doesn’t care about suffering if it comes from their pancreas. Thus, they’re in horrific misery, but it comes from their pancreas so they do nothing to prevent it, instead preventing a miniscule amount of non-pancreas agony. On anti-realism, they’ve made no error. But that’s crazy!
4 The Discovery Argument
One of the arguments made for mathematical platonism is the argument from mathematical discovery. The basic claim is as follows; we cannot make discoveries in purely fictional domains. If mathematics was invented not discovered, how in the world would we make mathematical discoveries? How would we learn new things about mathematics — things that we didn’t already know?
Well, when it comes to normative ethics, the same broad principle is true. If morality really were something that we made up rather than discovered, then it would be very unlikely that we’d be able to reach reflective equilibrium with our beliefs — wrap them up into some neat little web.
But as I’ve argued at great length, we can reach reflective equilibrium with our moral beliefs — they do converge. We can make significant moral discovery. The repugnant conclusion is a prime example of a significant moral discovery that we have made.
Thus, there are two facts about moral discovery that favor moral realism.
First, the fact that we can make significant numbers of non-trivial moral discoveries in the first place favors it — for it’s much more strongly predicted on the realist hypothesis than the anti-realist hypothesis.
Second, the fact that there’s a clear pattern to the moral convergence. Again, this is a hugely controversial thesis — and if you don’t think the arguments I’ve made in my 36-part series are at least mostly right, you won’t find this persuasive. However, if it turns out that every time we carefully reflect on a case it ends up being consistent with some simple pattern of decision-making, that really favors moral realism.
Consider every other domain in which the following features are true.
1 There is divergence prior to careful reflection.
2 There are persuasive arguments that would lead to convergence after adequate ideal reflection.
3 Many people think it’s a realist domain
All other cases which have those features end up being realist. This thus provides a potent inductive case that the same is true of moral realism.
5 The argument from phenomenal introspection
Credit to Neil Sinhababu for this argument.
If we have an accurate way of gaining knowledge and this method informs us of moral realism, then this gives us a good reason to be a moral realist, in much the same way that, if a magic 8 ball was always right, and it informed us of some fact, that would give us good reason to believe the fact.
Neil Sinhababu argues that we have a reliable way to gain access to a moral truth — this way is phenomenal introspection. Phenomenal introspection involves reflecting on a mental state and forming beliefs about what its like. Here are examples of several beliefs formed through phenomenal introspection.
My experience of the lemon is brighter than my experience of the endless void that I saw recently.
My experience of the car is louder than my experience of the crickets.
My experience of having my hand set on fire was painful.
We have solid evolutionary reason to expect phenomenal introspection to be reliable — after all, beings who are able to form reliable beliefs about their mental states are much more likely to survive and reproduce than ones that are not. We generally trust phenomenal introspection and have significant evidence for its reliability.
Thus, if we arrive at a belief through phenomenal introspection, we should trust it. Well, it turns out that through phenomenal introspection, we arrive at the belief that pleasure is good. When we reflect on what it’s like to, for example, eat tasty food, we conclude that it’s good. Thus, we are reliably informed of a moral fact.
Lance Bush has written a response to an article I wrote about this argument; I’ll address his response here.
I summarize Sinhababu’s argument as follows.
However, we can ignore premise one, because it serves as a reason other methods are unreliable — not as a reason phenomenal introspection is reliable. Lance says
I take a moral belief to be a belief about what is right and wrong, or what one should or shouldn’t do, or about what is good and bad. Morality is fundamentally about what we have impartial reason to do, independent of our desires. For more on this definition, I’d recommend reading Parfit’s On What Matters.
By reliable, I meant reliably true.
Lance here criticizes some types of introspection — however, none of this is phenomenal introspection. People are good at forming reliable beliefs about their experiences, less good at forming reliable beliefs about, for example, their emotions. Not all introspection is alike.
I think this objection to phenomenal conservatism is wrong. One can reject a seeming. For example, to me, the conclusion I describe here seems wrong, however, I end up accepting it upon reflection, because the balance of seemings supports it.
But we can table this discussion because Sinhababu doesn’t rely on seemings — he relies on phenomenal introspection.
No disagreement so far.
I agree that generally introspecting on experiences doesn’t inform us of their mind-independent goodness. But if we introspect on experiences that we don’t want but are pleasurable, they still feel good, showing that their goodness doesn’t depend on our desires.
But when you reflect on pleasure it feels good in a way that seems to give one a reason to promote it — to produce more of it. This is a distinctly moral notion. Sinhababu has a longer section on this in his paper — his account is somewhat different from mine.
Pleasure feels good in the sense that it’s desirable, worth aiming at, worth promoting. If this argument successfully establishes that pleasure is worth promoting, then it has done all that it needs to do. I don’t think morality is anything over and above a description of the things that are well and truly worth promoting.
This question is ambiguous, but I think the answer would be no.
I’d have a few things to say here.
1 It seems that most people have an intuitive sense of what it means to say something is wrong. This normal usage acquaintance is going to be more helpful than some formulaic definition that appears in a dictionary.
2 This seems rather like denying that there’s knowledge on the grounds that we don’t have a good definition of it. Things are very difficult to define — but that doesn’t mean we can’t be confident in our concepts of them. Nothing is ever satisfactorily defined.
3 I take morality to be about what we have impartial reason to aim at. In other words, what we’d aim at if we were fully rational and impartial.
Bush quotes me saying the following.
He responds.
The beliefs about what they’re like are beliefs about the experience. So, for example, the belief that hunger is uncomfortable is reliably formed through phenomenal introspection.
This was responded to above — when we reflect on pain we conclude that it’s the type of thing that’s worth avoiding, that there should be less of. We conclude this even in cases when we want pain. To give an example, I recall when I was very young wanting to be cold for some reason. I found that it still felt unpleasant, despite my desire to brave the cold.
Earlier in this article I was more precise and clarified the things that the anti-realist is committed to.
I think Lance does — he’s just terminologically confused. When he reflects on his pain, he concludes it’s worth avoiding — that’s why he avoids it! I think if he reflected on being in pain even in cases when he wanted to be in pain, he’d similarly conclude that it was undesirable.
6 Responding to Objections
A Disagreement
One common objection to moral realism is the argument from disagreement. The basic version is as follows.
Premise 1: If some domain has disagreement, then it only establishes subjective truths
Premise 2: The moral domain has disagreement
Therefore, it only establishes subjective truths
Problem: Premise 1 is obviously false. The domain of physics, mathematics, and numerous others garner lots of disagreement. They also are objective.
There are lots of more robust arguments from disagreement — however, I think the best paper on this subject by Enoch decisively refutes them.
B Access
Some worry about how we have access to the moral facts. Enoch puts these worries to rest decisively.
Enoch similarly describes why epistemic challenges for moral realism shouldn’t be thought of in terms of justification, reliability, or knowledge. I’d recommend the full paper for an explanation of this.
C Correlation
Enoch thinks the most puzzling version of epistemological objections don’t focus on any of the things above — instead, they focus on a puzzling correlation. This correlation is between the correct moral views and the moral things we happen to believe. Enoch says
On this, several points are worth making.
1 As Enoch points out, this is an explanatory game, so it makes sense to compare the explanatory adequacy of the theories holistically, and see if the best ones favor realism.
2 Also pointed out by Enoch, many people are in error, so the correlation isn’t that striking — it’s not as though there’s perfect correlation.
3 Our reasoning can weed out lots of views that are inconsistent — so that narrows the pool even more.
I’d also note
4 The correlation is not that striking — the correct moral view which seems to be hedonistic act utilitarianism is often wildly unintuitive.
5 Most of our beliefs tend to be right. Thus, based purely on priors, we’d expect the same broad pattern to be true when it comes to our moral beliefs.
6 The same broad arguments can be made against epistemic realism — it’s why there’s the correlation in that case too — but this doesn’t debunk our epistemic beliefs.
D Evolutionary Debunking
Street famously argued that our moral beliefs are evolutionarily debunkable — we formed them for evolutionary reasons, independent of their truth, so we shouldn’t believe them.
First, as Sinhababu points out, we’d expect evolution to make us reliable judges of our conscious experience. Belief in the badness of pain resists debunking because it’s formed through a mechanism that would evolve to be reliable. Much like beliefs about vision aren’t debunkable, neither are beliefs about our mental states, given that beings who can form accurate beliefs about their mental states are more likely to survive.
Second, as Bramble (2017) points out, evolution just requires that pain isn’t desired, it doesn’t require the moral belief that the world would be better if you didn’t suffer. Given this, there is no way to debunk normative beliefs about the badness of pain.
Third, there’s a problem of inverted qualia. As Hewitt (2008) notes, it seems eminently possible to imagine a being who sees red as blue and blue as red, without having much of a functional change. However, it seems like undesirability rigidly designates pain, such that you couldn’t have a being with an identical qualitative experience of pain, who seeks out and desires pain. This means that the badness and correlated undesiredness of pain is a necessary feature, not subject to evolutionary change.
One could object that there are many people like sadists who do, in fact, desire pain. However, when sadists are in pain, the experience they gain is one they find pleasurable. This is not a counterexample to the rule, so much as one that shows that experiences can have many features in common with pain, while lacking its intrinsic badness. A decent analogy here would be food--eating the same food at different times will produce different results, even with the same general taste. If one finds a food disgusting, their experience of eating it will be bad. Traditionally painful experiences are similar in this regard--closely related experiences can actually be desirable.
Fourth, evolution can’t debunk the direct acquaintance we have with the badness of pain, any more than it could debunk the belief that we’re conscious. Much like I have direct access to the fact that I’m conscious, I similarly have direct access to the badness of pain. After I stub my toe my conviction is much greater that the pain was bad than it is in the external world.
Fifth, it’s plausible that beings couldn’t be radically deluded about the quality of their hedonic experiences, in much the same way they can’t be deluded about whether or not they’re conscious. It seems hard to imagine an entity could have an experience of suffering but want more of it.
Sixth, there’s a problem of irreducible complexity. Pain only serves an evolutionary advantage if it’s not desired when experienced. Thus, the experience evolving by itself would do no good. Similarly, a mutation that makes a being not want to be in pain would do no good, unless it already feels pain. Both of those require the other one to be useful, so neither would be likely to emerge by themselves. However, only the intrinsic badness of pain which beings have direct acquaintance with can explain these two emerging together.
Seventh, evolution gave us the ability to do abstract, careful reasoning. This reasoning leads us to form beliefs about moral facts, in much the same way it does for mathematical facts.
E Explanatorily Unnecessary
People often object to moral realism on the grounds that the moral facts are explanatorily unnecessary. The earlier comments apply — positing real moral facts explains the convergence, for example, in our moral views. It also explains our moral seemings — seemings that inform us that, for example, it’s wrong to torture infants for fun and would be so even if nobody thought that it was.
F Objectionably Queer
Ever since the time of Mackie, it’s been objected that moral realism is objectionably queer, something about it is strange. However, it’s pretty unclear what exactly about it is supposed to be so strange. As Taylor says
However, it’s not clear why exactly this is so queer. As Huemer notes, many things are very different from everything else. Time is very different from other things, as is space, as are laws of physics — but we shouldn’t give up our belief in those things.
On top of this, it’s not clear why normativity is queer. There seem to be other things that are irreducibly normative — epistemic normativity seems on firm ground. One who believes the earth is flat on the basis of the available evidence is objectively making an epistemic error and, in an epistemic sense, they ought to change their views. None of this seems too queer.
Mackie just describes what morality is, before declaring that it’s too queer.
If you look at the attitudes of most everyday people towards the notion that it’s really wrong to torture infants for fun — it doesn’t seem strange at all to them.
Additionally, if one is too concerned about queerness, I think hedonism gives a particularly promising route for avoiding such worries. To quote my book.
Conclusion
Given the immense debate about moral realism, in this article, I have not been able to cover all of the relevant articles and arguments. However, I think I’ve summarized many of the main reasons to be a moral realist — some of which have, to the best of my knowledge, yet to be explored in the literature.
These arguments have been unapologetically pro hedonist. This is because I think that the challenges from the anti-realists to the hedonists are far weaker than they are for other moral realist views.
2 Non-physicalism about consciousness
0 A Brief Introduction
Why is there something rather than nothing? This question is quite difficult—perhaps even as difficult as the hard problem of consciousness. However, let’s consider some clearly terrible answers to the question.
There isn’t—something is an illusion.
Something is a weakly emergent property of nothing. When you have nothing for a little while, it combines to form something. Science will soon explain how nothing becomes something. Positing that there’s something that exists and is not reducible to nothing is like vitalism or phlogiston.
But, these answers are quite similar structurally and just as unsuccessful functionally as many “solutions,” to the hard problem of consciousness. In this blog post, I shall spell out why physicalist solutions to the hard problem fail—and why we need to be some type of dualist, idealist, or panpsychist.
Dispositionally, I’m an ardent physicalist. My intuitive, pre-theoretic leanings are strongly physicalist. However, when confronted with a brutal gang of facts, I was forced to abandon my physicalist leanings. This article draws heavily on the arguments of Chalmers in the conscious mind—definitely worth checking out for those who have not yet read it.
Let’s begin by defining physicalism. The SEP writes
1 Broad Considerations
“Consciousness is a biological phenomena,”
—John Searle, being wrong.
So why do I think that physical stuff cannot even in principle explain consciousness. Well, there are two closely related higher order considerations, and then some more specific arguments.
The first broad consideration which explains why consciousness resists physicalist reduction is that physics explains things in terms of structure and function, as Chalmers notes. Physics gives equations to describe what things do and what they’re composed of. However, this cannot in principle explain what it’s like to eat a strawberry, see the color red, be in love. When we look at an atom, we have no way of verifying whether or not it is conscious, because we only observe its causal impacts.
So this is not analogous to Phlogiston or vitalism or anything else physicalists use as an analogy for consciousness. All of those are broadly explicable in terms of structure and function, and thus they don’t require any extra laws. Consciousness is different—it’s not in principle explainable in terms of structure and function.
A second related broad consideration which has been expressed eloquently by Kastrup is that material stuff can be exhaustively explained quantitatively. Through physics, we get a series of equations. To quote Kastrup
2 Zombies
—David Chalmers
An adequate model of physics will be able to describe what physically goes on in your brain. However, we can imagine a physical carbon copy of you that lacks consciousness. This shows consciousness is not purely physical, as we can’t imagine a carbon copy of H20 that isn’t water.
One confusion had by many is that the zombie argument presumes some type of epiphenomenalism, the notion that consciousness has no physical effect. This is false. If consciousness has a physical effect, the zombie would have some other law of physics fill in and play the functional role of consciousness. So if consciousness causes me to say things like “I’m conscious,” “I think therefore I am,” “consciousness poses a hard problem,” “Dan Dennett might be a zombie,” “consciousness can’t be explained reductively,” “Okay—Dennett is definitely a zombie,” etc—the zombie world would have some physically identical force fill in the functional role of consciousness and cause me to say all of those things.
Thus, the argument is as follows.
1 A being could be physically identical to me but could not be conscious
2 Two beings that are physically identical must have all physical properties in common
Therefore, consciousness is not a physical property.
There’s much more that can be said on the topic of zombies, however, to me it seems quite obvious that zombies are possible—those who deny their possibility seem conceptually confused to be. No doubt that’s how I seem to them. Yet I haven’t the time in this article to go into all of the accounts of the alleged impossibility of zombies, yet it’s worth noting that zombies are somewhat controversial.
3 Inverted Qualia
—Chalmers.
If consciousness just is a physical phenomena, then it would be impossible to change conscious experiences without making a physical change. However, it seems eminently metaphysically possible that we could change consciousness but not make a physical change. Imagine a world physically identical to ours but in which one tomato that I see appears 1% redder than it does currently. If you think that world is possible, then consciousness is not purely physical.
Note, I’m perfectly willing to grant that based on the world as it currently exists, such a state would be impossible. There are, in my view, psychophysical laws that govern consciousness which make it so that consciousness can’t be different. However, we could make tweaks to those laws without having a physical effect, which shows consciousness is not physical.
4 Epistemic Asymmetry
You guessed it.
If consciousness were a reductively explainable physical property, then we’d be able to deduce its existence from knowledge of the lower level facts. However, this is manifestly impossible in the case of consciousness. If you knew everything abut atoms, you’d be able to deduce the existence of fire and explain what it does. However, nothing about consciousness is evident from low level descriptions of physical systems.
Why do you think others are conscious? Well, the reason is because you know you’re conscious and others plausibly have similar features to the ones that make you conscious. However, this is not how we deduce that others can get sick. Rather, we directly observe others getting sick. Even if we were in perfect health, it would be reasonable to infer that others get sick. However, if you were not conscious, it would not be reasonable to infer others were conscious. This is because consciousness is not explainable by low level physical facts.
5 The Knowledge Argument
—Guess who!
If consciousness were a reductively explainable physical property then knowing all of the facts about the brain would make it possible to know what it’s like to see red, despite being color blind. However, this is clearly impossible. No neuroscientific knowledge can communicate what it’s like to see red, for one who has never seen red. If Mary left the room and saw a red tomato, she’d learn something new about what it’s like to see red. Her curiosity would be satisfied by seeing the color red, if she had previously wondered what it was like to see red.
No amount of neurological knowledge could teach a deaf person what it’s like to hear Mozart or a blind person what it’s like to see the grand canyon. However, if consciousness were purely physical, this would be possible. If one knows all of the facts about bricks, they could know all relevant facts about brick walls. This is because a brick wall is an emergent property of bricks. If consciousness were merely physical, then much like full physical knowledge would teach you everything there is to be known about a tumor, supernova, or ocean, the same would be true of consciousness. However, this is manifestly impossible.
6 From The Absence Of Analysis
—Greg, just kidding, Chalmers obviously.
When we consider facts about a physical system, none of them make it obvious why those things would make it conscious. Consider, for example, the integrated information theory, which says that when one system processes a variety of different types of information, it becomes conscious, with its consciousness proportional to the amount of integrated information. When information is integrated, nothing about that physical state obviously produces consciousness. It seems like there’s a further question—we know a system has integrated information, but that doesn’t settle whether it’s conscious.
Consciousness is not just integrated information. It seems imminently possible to imagine a non conscious system that integrates information. When we identify the neural correlates of consciousness, it’s never obvious why those things would be conscious. We can understand why H20 is water, but no explanation of why the neural correlates of consciousness are consciousness.
7 Disembodied Minds
If consciousness were just a physical phenomena, then disembodied minds would be metaphysically impossible. Because heat just is the rapid movements of particles, disembodied heat is impossible. To have heat, one needs particles to move rapidly.
It would make no sense to talk about a non-physical tortoise, box, or pancreas, because these are physical phenomena. However, disembodied minds—minds without bodies—seem metaphysically possible. We could imagine mental functions going on, even in the absence of a body. This shows that consciousness isn’t a purely physical property—it could exist in the absence of physical things.
8 Some Concluding Thoughts On Why This Isn’t Vitalism
Vitalism is the notion that living organisms have some fundamental life causing non-physical substance—“Élan vital.” Many have given analogies between non-physicalism about consciousness and vitalism, as they both posit a non material thing. However, it’s worth noting that none of the arguments above can apply to vitalism.
Life just is about structure and function and can be described quantitatively—so it’s not susceptible to the first argument. A L zombie, physically identical to an alive thing but that isn’t alive, is obviously impossible. It is possible to use low level phenomena to explain life, unlike for consciousness. There’s no analogy for the inverted qualia argument. Knowing all the physical facts about a physical system would let you know whether it’s alive and all the facts about its life, there is an account of how cells replicate and comprise life, and disembodied life is obviously impossible.
The properties that were appealed to for vitalism were non-physical properties, but ones that we now know don’t exist. There’s nothing it is to be alive over and above the physical facts relating to cell replication, growth, and the other things required for life. Thus, the correct view about vitalism was illusionism—the properties being posited that Elan Vital explained weren’t real. But we know consciousness is real! It’s the most certainly known natural phenomena—we can be more certain that we’re conscious than we can be of anything else.
Abandoning physicalism isn’t abandoning an answer to the problems of consciousness—it merely recognizes the reality of what form the answer must take. Non physicalist theories are testable and make predictions which can be subsequently verified.
Sometimes, the correct answers are surprising and run afoul of our heuristics. Generally people worrying about new technology are wrong, but not when it comes to AI alignment. Usually, Parfit is right, but not when it comes to the repugnant conclusion. Preachy vegans are irritating, but they’re right. Reductionism is enticing—it would be so nice if consciousness were just some physical phenomena, but there are knockdown arguments against such a view. We mustn’t be held captive to reductionist dogma, in the face of overwhelming evidence.
Eliezer is provably wrong about zombies
I enjoy much of what Eliezer Yudkowsky says. He’s been a large part of raising worries about AI alignment, writes tons of interesting less wrong stuff, wrote the epic HPMOR, and has shaped my thinking in many ways. However, Yudkowsky is, as the title hints at, wrong about zombies.
A zombie is a being physically identical to a conscious being in every way, minus the consciousness. The important thing to note is that the zombie would, if consciousness is causally efficacious, have other things that fill in the causal roles of the person.
Yudkowsky writes
Note, when we use possibility here, we’re describing metaphysical possibility, not physical possibility. So the question is whether there is a possible world that is atom for atom identical to this world but that lacks consciousness. All of the things done by consciousness would be done by other laws that are functionally identical to consciousness in this world, but that don’t contain any experiences.
Eliezer’s claim that this view is epiphenomenalism is false. Epiphenomenalism says consciousness doesn’t cause anything. One can hold to epiphenomenalism and zombies, because the zombie world would have something else do what your consciousness does in this world.
But it is a strawman!! The zombie argument doesn’t entail epiphenomenalism. It’s often made by interactionist dualists, panpsychists, and idealists. It’s frustrating that Eliezer strawman’s the arguments while specifically talking about not straw manning it. I’m not suggesting bad faith here, it’s just a bit frustrating.
If we look inside the brain, what we see happening involves the flow of electric signals from your brain to the muscles in your arm, resulting in the refrigerator opening. The point of the zombie argument is that you could imagine a world where all of that goes on in exactly the same way—it looks precisely the same from the outside in terms of the movement of all of the atoms—but you are not conscious when it goes on.
I’m not an epiphenomenalist (my credence in it is around 10%), but the epiphenomenalists can give an explanation of this. If consciousness is just what it feels like for the brain to do things, then it feels like you’re the cause of it, but really your consciousness just is what it feels like for the brain to do things.
This is false. When Chalmers is defining views about philosophy of mind, he writes
Obviously epiphenomenalism is different from Descartes’ dualism. Descartes was a substance dualist and interactionist. These extra views aren’t required for dualism. Zombieism, as Eliezer calls it, can be dualist or panpsychist—it just has to reject physicalism.
This is of course true. The question for zombies isn’t just whether we could imagine them—I could imagine fermat’s last theorem being false, but it isn’t—but whether it’s metaphysically possible that they exist.
I think this is a pretty good argument against epiphenomenalism. However, this does nothing to show that consciousness is physical, and it doesn’t answer the zombie argument. Consider an analogy—imagine that the cause of gravity is a god willing gravity to be so, one who is defined as being non-physical. Even though gravity is caused by the non-physical mind, we could imagine a world that’s physically identical, where gravity is caused by something else, other than the non-physical mind. Consciousness is the same.
But again, you could have some functional analogue that does the same physical thing that your consciousness does. Any physical affect that consciousness has on the world could be in theory caused by something else. If consciousness has an affect on the physical world, it’s no coincidence that a copy of consciousness would have to be hyper specific and cause you to talk about consciousness in exactly the same way.
This is not true. As this paper notes
Chalmers himself notes in a comment below the original post
Thus, even if consciousness causes things, that’s just a description of what consciousness does. One could imagine a world where all the atoms move in the same way, as if they’re prompted by consciousness, but they aren’t caused by anything conscious. A subjective experience may do something causally, but you could imagine a physical law on interactionism that does exactly the same things consciousness does. Next Eliezer says
Interactionism doesn’t hold that consciousness is not experimentally detectable—that’s not a necessary entailment of dualism. The zombie world on interactionism wouldn’t need an extra zombie master. Suppose that the psychophysical law in this world is that when you get a bunch of neurons together they become conscious and then their desires exert some force. Well, the zombie world would have the same forces exerted, just minus the mental state of desires.
The reason is because consciousness is not merely causal. It does cause things, but there’s something it’s like to see red, over and above what it causes. Thus, it’s possible that you could take away that other stuff in theory, and still have a causal isomorph. The reason people postulate that consciousness is causally inert is because
A) there are problems incorporating its causal role into physics.
B) All one has to posit is that when a person has a particular desire, that corresponds with the physical effect. Epiphenomenalists argue that the simplest consciousness laws involve the physical state that is about to raise your arm causing consciousness, rather than the other way around.
One can be an interactionist property dualist. Property dualism just requires saying that consciousness is a property of matter, not its own separate substance.
I lean towards interactionist dualism, so I’m in agreement with Eliezer here. However, the claim that dualism is motivated by finding something that seems mysterious and then just positing mysterious stuff is totally wrong. Dualists don’t just give up on explanations—there are lots of ways that specific dualists have experimentally tested their theories.
There are lots of reasons to posit dualism of some sort, which I lay out here. The fundamental reason is that the laws of physics explain physics in terms of structure and function—yet none of that is able to explain the subjective experience of seeing red, for example. Subjective experience is neither structural nor functional, so the physics based account that explains it in terms of structure and function is wholly inadequate.
This is not accurate. For one, Chalmers is now pretty undecided between different versions of non-physicalism. Chalmers objects to substance dualism based on it violating causal closure of the physical, having trouble explaining how consciousness would interact, and plausibly being ruled out by physics.
Overall, I quite like Eliezer, as I said at the outset. However, it’s frustrating that when it comes to consciousness, he just seems very lost. This is particularly a problem given that consciousness is literally the most important thing in the universe—the only important thing in the universe. So it’s really, really, really important not to get things wrong, when it comes to consciousness.
Eliezer at one point says
Yet within physics, the last 3000 times, we haven’t just posited the same old laws. Newton discovered brand new laws, so did Einstein. Consciousness is not more fundamentally mysterious—there are just some type of fundamental psychophysical laws that result in consciousness, which have a causal effect on the world.
Trying to explain it with the same old stuff, when we have lots of knock-down arguments against the ability of the old stuff to explain it—that’s an appeal to magic. Eliezer’s reductive account involves positing that when you have some physical things, they just produce experience, despite our inability to either
A) Understand how physics could go beyond explaining structure and function.
B) Provide any account of how brain stuff generates consciousness.
C) Provide a physical description of any type of conscious state.
All of the other accounts of successful reduction have involved explaining the behavior of things at a higher level, by appealing to lower level facts. But this just won’t work for consciousness! Consciousness isn’t about behavior. When we ask whether AI is conscious, we don’t care whether they say verbally that they’re conscious. What we care about is whether the ineffable what it’s like stuff is present in the AI.
This is a seriously important mistake for effective altruists not to make. We must not with away and ignore the fundamental difficulty of the hardest problem in the universe. Saying “it just emerges,” is not a good solution. And yet I fear that’s the solution of many of my fellow effective altruists and rationalists—a mistake that could be very costly.