I find negative utilitarianism unappealing for roughly the same reason I'd find "we should only care about disgust" or "we should only care about the taste of bananas" unappealing. Or if you think suffering is much closer to a natural kind than disgust, then supply some other mental (or physical!) state that seems more natural-kind-ish to you.
"Only suffering ultimately matters" and "only the taste of bananas ultimately matters" share the virtue of simplicity, but they otherwise run into the same difficulty, which is just that they don't exhaustively describe all the things I enjoy or want or prefer. I don't think my rejection of bananatarianism has to be any more complicated than that.
Something I wrote last year in response to a tangentially related paper:
I personally care about things other than suffering. What are negative utilitarians saying about that?
Are they saying that they don't care about things like friendship, good food, joy, catharsis, adventure, learning new things, falling in love, etc., except as mechanisms for avoiding suffering? Are they saying that I'm deluded about having preferences like those? Are they saying that I should try to change my preferences — and if so, why? Are they saying that my preferences are fine in my personal decision-making as an individual, but shouldn't get any weight in an idealized negotiation about what humanity as a group should do (ignoring any weight my preferences get from non-NU views that might in fact warrant a place at the bargaining table for more foundational or practical reasons distinct from the NU ideal) — and if so, why?
[...] "It's wrong to ever base any decision whatsoever on my (or anyone else's) enjoyment of anything whatsoever in life, except insofar as that enjoyment has downstream effects on other things" is an incredibly, amazingly strong claim. And it's important in this context that you're actually making that incredibly strong claim: more mild "negative-leaning" utilitarianisms (which probably shouldn't be associated with NU, given how stark the difference is) don't have to deal with the version of the world destruction argument I think x-risk people tend to be concerned about, which is not 'in some scenarios, careful weighing of the costs and benefits can justify killing lots of people' but rather 'any offsets or alternatives to building misaligned resource-hungry AGI (without suffering subsystems) get literally zero weight, if you're sufficiently confident that that's what you're building; there's no need to even consider them; they aren't even a feather on the scale'. I just don't see why the not-even-a-feather-on-the-scale view deserves any more attention or respect than, e.g., divine-command theory; in an argument between the "negative-leaning" utilitarian and the real negative utilitarian, I don't think the NU gets any good hits in.
(Simplicity is a virtue, but not when it's of the "I'm going to attempt to disregard every consideration in all of my actions going forward except the expected amount of deliciousness in the future" or "... except the expected amount of lying in the future" variety; so simplicity on its own doesn't raise the view to the level of having non-negligible probability compared to negative-learning U.)
Yes, I am making the (AFAICT, in your perspective) “incredibly, amazingly strong claim” that in a unified theory, only suffering ultimately matters. In other words, impartial compassion is the ultimate scale (comparator) to decide conflicts between expected suffering vs. other values (whose common basis for this comparison derives from their complete, often context-dependent relationship to expected suffering, including accounting for the wider incentives & long-term consequences from breaking rules that are practically always honored).
I find negative...
In evolutionary and developmental history terms, we can see at the first quick glance that many (if not immediately all) of our other motivations interact with suffering, or have interacted with our suffering in the past (individually, neurally, culturally, evolutionarily). They serve functions of group cohesion, coping with stress, acquiring resources, intimacy, adaptive learning & growth, social deterrence, self-protection, understanding ourselves, and various other things we value & honor because they make life easier or interesting.
Seems like all of this could also be said of things like "preferences", "enjoyment", "satisfaction", "feelings of correctness", "attention", "awareness", "imagination", "social modeling", "surprise", "planning", "coordination", "memory", "variety", "novelty", and many other things.
"Preferences" in particular seems like an obvious candidate for 'thing to reduce morality to'; what's your argument for only basing our decisions on dispreference or displeasure and ignoring positive pre...
This comment doesn't seem to sufficiently engage with (what I saw as) the core question Rob was asking (and which I would ask), which was:
I personally care about things other than suffering. What are negative utilitarians saying about that?
Are they saying that they don't care about things like friendship, good food, joy, catharsis, adventure, learning new things, falling in love, etc., except as mechanisms for avoiding suffering? Are they saying that I'm deluded about having preferences like those? Are they saying that I should try to change my preferences — and if so, why?
You briefly note "you may be overly attached to them", but this doesn't give any arguments for why I might be overly attached to them, instead of attached to them the correct amount.
When you ask:
To actually reject NU, you must explain what makes something (other than suffering) terminally valuable (or as I say, motivating) beyond its instrumental value for helping us prevent suffering in the total context.
My response is "to reject NU, all I have to do is terminally care about anything other than suffering. I care about things other than suffering, ergo NU must be false, and the burden is on other people to explain what is wrong with my preferences."
I used to consider myself NU, but have since then rejected it.
Part of my rejection was that, on a psychological level, it simply didn't work for me. The notion that everything only has value to the extent that it reduces suffering meant that most of the things which I cared about, were pointless and meaningless except for their instrumental value in reducing my suffering or making me more effective at reducing suffering. Doing things which I enjoyed, but constantly having a nagging sensation of "if I could just learn to no longer need this, then it would be better for everyone" basically meant that it was very hard to ever enjoy anything. It was basically setting my mind up to be a battlefield, dominated by an NU faction trying to suppress any desires which did not directly contribute to reducing suffering, and opposed by an anti-NU faction which couldn't do much but could at least prevent me from getting any effective NU work done, either.
Eventually it became obvious that even from an NU perspective, it would be better for me to stop endorsing NU, since that way I might end up actually accomplishing more suffering reduction than if I continued to endorse NU. And I think that this decision was basically correct.
A related reason is that I also rejected the need for a unified theory of value. I still think that if you wanted to reduce human values into a unified framework, then something like NU would be one of the simplest and least paradoxical answers. But eventually I concluded that any simple unified theory of value is likely to be wrong, and also not particularly useful for guiding practical decision-making. I've written more about this here.
Finally, and as a more recent development, I notice that NU neglects to take into account non-suffering-based preferences. My current model of minds and suffering is that minds are composed of many different subagents with differing goals; suffering is the result of the result of different subagents being in conflict (e.g. if one subagent wants to push through a particular global belief update, which another subagent does not wish to accept).
This means that I could imagine an advanced version of myself who had gotten rid of all personal suffering, but was still motivated by pursue other goals. Suppose for the sake of argument that I only had subagents which cared about 1) seeing friends 2) making art. Now if my subagents reached agreement of spending 30% of their time making art and 70% of their time seeing friends, then this could in principle eliminate my suffering by removing subagent conflict, but it would still be driving me to do things for reasons other than reducing suffering. Thus the argument that suffering is the only source of value fails; the version of me which had eliminated all personal suffering might be more driven to do things than the current one! (since subagent conflict was no longer blocking action in any situation)
As a practical matter, I still think that reducing suffering is one of the most urgent EA priorities: as long as death and extreme suffering exist in the world, anything that would be called "altruism" should focus its efforts on reducing that. But this is a form of prioritarianism, not NU. I do not endorse NU's prescription that an entirely dead world would be equally good or better as a world with lots of happy entities, simply because there are subagents within me who would prefer to exist and continue to do stuff, and also for other people to continue to exist and do stuff if they so prefer. I want us to liberate people's minds from involuntary suffering, and then to let people do whatever they still want to do when suffering is a thing that people experience only voluntarily.
I think most of this is compatible with preference utilitarianism (or consequentialism generally), which, in my view, is naturally negative. Nonnegative preference utilitarianism would hold that it could be good to induce preferences in others just to satisfy these preferences, which seems pretty silly.
Thanks for the perspective.
I agree that even NU may imply rejecting NU in its present form, because it does not feel like a psychologically realistic theory to constantly apply in everyday life; we are more motivated to move towards goals and subgoals that do not carry explicit reminders of extreme suffering on the flip side.
I do feel that I am very close to NU whenever I consider theoretical problems and edge-cases that would outweigh extreme suffering with anything else than preventing more extreme suffering. In practice, it may be more applicable (and i...
I believe the most often cited (in the LW/EA communities) paper arguing against NU is Toby Ord's Why I'm Not a Negative Utilitarian. This and this seem to be the main replies to it from NU perspectives. (I think I've skimmed some of these articles but have not actually considered the arguments carefully.)
Ethical theories don't need to be simple. I used to have the belief that ethical theories ought to be simple/elegant/non-arbitrary for us to have a shot at them being the correct theory, a theory that intelligent civilizations with different evolutionary histories would all converge on. This made me think that NU might be that correct theory. Now I’m confident that this sort of thinking was confused: I think there is no reason to expect that intelligent civilizations with different evolutionary histories would converge on the same values, or that there is one correct set of ethics that they "should" converge on if they were approaching the matter "correctly". So, looking back, my older intuition feels confused now in a similar way as ordering the simplest food in a restaurant in expectation of anticipating what others would order if they also thought that the goal was that everyone orders the same thing. Now I just want to order the "food" that satisfies my personal criteria (and these criteria do happen to include placing value on non-arbitrariness/simplicity/elegance, but I’m a bit less single-minded about it).
Your way of unifying psychological motivations down to suffering reduction is an "externalist" account of why decisions are made, which is different from the internal story people tell themselves. Why think all people who tell different stories are mistaken about their own reasons? The point "it is a straw man argument that NUs don’t value life or positive states“ is unconvincing, as others have already pointed out. I actually share your view that a lot of things people do might in some way trace back to a motivating quality in feelings of dissatisfaction, but (1) there are exceptions to that (e.g., sometimes I do things on auto-pilot and not out of an internal sense of urgency/need, and sometimes I feel agenty and do things in the world to achieve my reflected life goals rather than tend to my own momentary well-being), and (2) that doesn’t mean that whichever parts of our minds we most identify with need to accept suffering reduction as the ultimate justification of their actions. For instance, let’s say you could prove that a true proximate cause why a person refused to enter Nozick’s experience machine was that, when they contemplated the decision, they felt really bad about the prospect of learning that their own life goals are shallower and more self-centered than they would have thought, and *therefore* they refuse the offer. Your account would say: "They made this choice driven by the avoidance of bad feelings, which just shows that ultimately they should accept the offer, or choose whichever offer reduces more suffering all-things-considered.“ Okay yeah, that's one story to tell. But the person in question tells herself the story that she made this choice because she has strong aspirations about what type of person she wants to be. Why would your externally-imported justification be more valid (for this person's life) than her own internal justification?
Thanks for the replies, everyone!
I don’t have the time to reply back individually, but I read them all and believe these to be pretty representative of the wider community’s reasons to reject NU as well.
I can’t speak for those who identify strictly as NU, but while I currently share many of NU’s answers to theoretical outweighing scenarios, I do find it difficult to unpack all the nuance it would take to reconcile “NU as CEV” with our everyday experience.
Therefore, I’ll likely update further away from
{attempting to salvage NU’s reputation by bridging it with compassion, motivation theory, and secular Buddhism}
towards
{integrating these independent of NU, seeing if this would result in a more relatable language, or if my preferred kind of theoretical unity (without pluralist outweighing) would still have the cost of its sounding absurd and extreme on its face}
Did you make any update regarding the simplicity / complexity of value?
My impression is that theoretical simplicity is a major driver of your preference for NU, and also that if others (such as myself) weighed theoretical simplicity more highly that they would likely be more inclined towards NU.
In other words, I think theoretical simplicity may be a double crux in the disagreements here about NU. Would you agree with that?
If the rationality and EA communities are looking for a unified theory of value
Are they? Many of us seem to have accepted that our values are complex.
Absolute negative utilitarianism (ANU) is a minority view despite the theoretical advantages of terminal value monism (suffering is the only thing that motivates us “by itself”) over pluralism (there are many such things). Notably, ANU doesn’t require solving value incommensurability, because all other values can be instrumentally evaluated by their relationship to the suffering of sentient beings, using only one terminal value-grounded common currency for everything.
This seems like an argument that it would be convenient if our values were simple. This does not seem like strong evidence that they actually are simple. (Though I grant that you could make an argument that it might be better to try to achieve only part of what we value if we're much more likely to be successful that way.)
What have you read about it that has caused you to stop considering it, or to overlook it from the start?
I reject impartiality on the grounds that I'm a personal identity and therefore not impartial. The utility of others is not my utility, therefore I am not a utilitarian. I reject unconditional altruism in general for this reason. It amazes me in hindsight that I was ever dumb enough to think otherwise.
Can you teach me how to see positive states as terminally (and not just instrumentally) valuable, if I currently don’t?
Teach, no, but there are some intuitions that can be evoked. I'd personally take a 10:1 ratio between pleasure and pain; if I get 10 times more pleasure out of something, I'll take any pain as a cost. It's just usually not realistic, which is why I don't agree that life has generally positive value.
There are fictional descriptions of extreme pleasure enhancement and wireheading, e.g. in fantasy that describe worthwhile states of experience. The EA movement is fighting against wireheading, as you can see in avturchin's posts. But I think such a combination of enhancement + wireheading could plausibly come closest to delivering net-positive value of life, if it could be invented (although I don't expect it in my lifetime, so it's only theoretical). Here's an example from fiction:
"You see, I have a very special spell called the Glow. It looks like this." The mage flicked his fingers and they started Glowing in a warm, radiant light. Little Joanna looked at them in awe. "It's so pretty! Can you teach me that? Is that the reward?" Melchias laughed. "No. The true reward happens when I touch you with it." She stared at him curiously. He looked down at the table in front of him with an expectant look, and she put her slender arm there so he could touch her. "Here, let me demonstrate. Now, this won't hurt one bit..." He reached out with his Glowing fingers to touch the back of her small hand, ever so gently.
And as their skin connected, Joanna's entire world exploded. The experience was indescribable. Nothing, no words and no warnings, could have prepared Joanna for the feeling that was now blasting through her young mind with the screaming ferocity of a white-hot firestorm, ripping her conscious thoughts apart like a small, flickering candlelight in a gigantic hurricane of whipping flames and shredding embers. She had no idea, no idea such pleasure existed, ever could exist! It was all of her happy memories, all of her laughter, her playfulness, her sexy tingles when she rubbed herself between the legs, the goodness of eating apple pie, the warmth of the fireplace in the winter nights, the love in her papa's strong arms, the fun of her games and friendships with the other village kids, the excitement of stories yet unheard, the exhilaration of running under the summer sun, the fascination of the nature and the animals around her, the smells and the tastes, the hugs and awkward kisses, all the goodness of all her young life, all condensed into a mere thousandth of a split-second... ...and amplified a thousand-fold... ...and shot through her mind, through her soul, again and again and again, split-second after split-second after split-second, like a bombardment of one supernova after another supernova of pure, unimaginable bliss, again and again and again, and yet again, second after second, filling her up, ripping her apart with raw ecstatic brilliance, melting her mind together in a new form, widened and brighter than it had ever been, a new, divine, god-like Joanna that no words could adequately worship, only to rip her apart again with a new fiery pulse of condensed, sizzling-hot vibrance, indescribable, unimaginable, each second an unreached peak, a new high, a new universe of fantastic pleasure, a new, unspeakably wonderful Joanna, loved and pulsing in her own Glowing light with a beauty unmatched by any other thing in all of the World Tree. She was a giant beating heart that was also a Goddess, Glowing and pulsing in the center of Everything, Glowing with the certainty of absolute affirmation, the purity of absolute perfection, the undeniability of absolute goodness. She spent centuries in seconds, serene yet in screaming ecstasy, non-living yet pulsing raw life force, non-personal yet always Joanna, Joanna in divine totality. It took Joanna a long time to realize she was still breathing, a living child with an actual human body. She had forgotton to breathe, and was sure she would have suffocated by now, but somehow, inexplicably, her body had remembered to keep itself alive without her. Master Melchias had lied: It did hurt, her chin and lip hurt, but the young girl found it was only because she had dropped to the hard stone floor in helpless twitching convulsions, and she had accidentally bitten herself. As promised by the wizard, the small wound was quickly healing. Joanna couldn't get up yet. She had no idea how much time had passed, but she just couldn't move or even open her young eyes yet. She curled up into a fetal position on the cold, hard floor of Melchias' Tower and sobbed uncontrollably. She sobbed and cried, and sobbed, and laughed madly, then sobbed and cried again. They were tears of pure joy.
The Glow wasn't just a normal pleasure spike, like an orgasm, a fit of laughter or a drug high. It went far, far beyond that. Normal human experiences existed within an intensity range that was given by nature. It served to motivate the organism for survival and reproduction, but it was not optimized for the experience itself. Even the most intense experiences, like burning alive or being skinned alive, existed within that ordinary, natural range. But the magic of the Glow didn't just stimulate pleasure within that range - it completely changed the range itself. It broke the scale on which normal experiences were measured, and then attached a vast multitude of additional ones to its top. By enhancing the part of the subject's mind that contained ordinary pleasure, it became temporarily able to experience an intensity that was hundreds of thousands times stronger than even the most extreme natural human feeling. Being drowned in hot oil, being flayed alive or tortured with needles, deep romantic love and fulfillment, orgasmic ecstasy, perfect fits of laughter - all of these human extremes represented only a miniscule fraction of the new potential. And only then did the spell induce raw, optimized pleasure within this new, widened consciousness. The result was an unimaginably pure goodness that fell so far outside of the subject's prior experience that it couldn't even be communicated by words. It had to be demonstrated. Once a potential [...] candidate had perceived even one second of the Glow, each containing more joy and happiness than an average human lifespan, with none of its pain, they all became devoted followers to Melchias. He transformed their experience from something human to something divine, and in turn, he became like a god to them.
The utility of others is not my utility, therefore I am not a utilitarian. I reject unconditional altruism in general for this reason.
When I say that I'm a utilitarian (or something utilitarian-ish), I mean something like: If there were no non-obvious bad side-effects — e.g., it doesn't damage my ability to have ordinary human relationships in a way that ends up burning more value than it creates — I'd take a pill that would bind my future self to be unwilling to sacrifice two strangers to save a friend (or to save myself), all else being equal.
The not-obviously-confused-or-silly version of utilitarianism is "not reflectively endorsing extreme partiality toward yourself or your friends relative to strangers," rather than "I literally have no goals or preferences or affection for anything other than perfectly unbiased maximization of everyone's welfare'".
If you flip the Rachels-Temkin spectrum argument (philpapers.org/archive/NEBTGT.pdf), then some tradeoff between happiness and suffering is needed to keep transitive preferences, which is necessary to avoid weird conclusions like accepting suffering to avoid happiness. As long as you don't think theres some suffering threshold where 1 more util of suffering is infinitely worse than anything else, then this makes sense.
Also NU in general has a bad reputation in the philosophy community (more than classical utilitarianism I think) so it's better EAs don't endorse it.
If you flip the Rachels-Temkin spectrum argument (philpapers.org/archive/NEBTGT.pdf), then some tradeoff between happiness and suffering is needed to keep transitive preferences, which is necessary to avoid weird conclusions like accepting suffering to avoid happiness. As long as you don't think theres some suffering threshold where 1 more util of suffering is infinitely worse than anything else, then this makes sense.
Can you give a practical example of a situation where I would be hereby forced to admit that happiness has terminal value above its ins...
Meta-note: I am surprised by the current karma rating of this question. At present, it is sitting at +9 points with 7 votes, but it would be at +2 with 6 votes w/o my strong upvote.
To those who downvoted, or do not feel inclined to upvote -- does this question not seem like a good use of LW's question system? To me it seems entirely on-topic, and very much the kind of thing I would want to see here. I found myself disagreeing with much of the text, but it seemed to be an honest question, sincerely asked.
Was it something about the wording (either of the headline or the explanatory text) that put you off?
I didn't vote - it seems appropriate for LW, but didn't have any links or references, so it seemed to be a reaction to an argument that I wasn't familiar with. The bold text (only one line, so presumably the main point of the post) was dismissing something as a strawman without explaining why.
It did spawn some interesting conversation, so I've now upvoted it.
Serious question: are you depressed? So far most negative utilitarians I've known were depressed and some stopped being negative utilitarian after fixing a chemical imbalance that hampered their ability to experience good things.
No, I’m not depressed, and I believe I never have been. I understand and appreciate the question if what you describe is your prior experience of people who identify as negative utilitarians. I may identify as NU for discussion’s sake, but my underlying identification is with the motivation of impartial compassion. I would go as far as to say that I am happy in all areas of my personal life, being driven towards unification by my terminal concern for the expected suffering of others.
I have had brief experiences of medical emergencies that gave me new perspectives into suffering from the inside. (In hindsight, much of it was generated by fear and escalating perception, but so it is in real danger.) While those happened years ago, I’ve continued to reflect on them and feel like they’ve changed me, affecting my daily life and priorities since then. For a while, I felt grateful for getting my life back and considered devoting myself to treating acute pain. I since graduated Master’s in Psychology without clinical internship to focus more on research, feeling that my comparative advantage is in channeling compassion for more scalable, theoretical work.
I believe a possible mistake of depressed NUs is to focus on others’ suffering before taking care of themselves (by listening to the foundational motivation of self-compassion). Self-compassion is our value-grounding for extended self-compassion, which leads to universal compassion in the limit.
Nate Soares has a post about self-compassion as a key part of his wider, 40-post series, Replacing Guilt, both of which I universally recommend (also ePUB-compiled here).
I like tranquilism as a view preserving some of the attractive aspects of negative utilitarianism whilst sidestepping some of its icky implications.
Can you provide a pointer or summary to how ANU doesn't require commensurability of values? How does one evaluate reduction in suffering of many small-experience/short-lived creatures at the expense of increased suffering of a deeper-experiencing/longer-lived creature?
Also, how does ANU (or any system trying to minimize it's only metric) deal with population-size issues? Should Thanos snap his fingers a lot more often (removing suffering by removing entities, painlessly and by surprise)?
I'll admit that neither of these are my true objection. My felt experience and strong intuition is that suffering can coexist with joy and satisfaction, and one should not sacrifice much of the latter in order to reduce the former.
I haven't seriously evaluated the arguments, but my intuition is that suffering and happiness are opposite sides of the same scale, not separate values. Utility is the measure of how good or bad something is, and happiness and suffering correspond to positive and negative values of utility.
terminal value monism (suffering is the only thing that motivates us “by itself”)
So I'd say that I value only utility.
it is a straw man argument that NUs don’t value life or positive states, because NUs value them instrumentally
Again, not having thought too much about it, I find that my intuition better matches a system that cares about positive utility even when it doesn't avert negative utility. E.g., I want paradise forever, not mild pleasantness forever.
Is there a good reason to suspect this is wrong?
Note (2022-03-15): For a recent defense of similar axiological views, see my series on minimalist axiologies.
(Crossposted on the EA Forum)
Absolute negative utilitarianism (ANU) is a minority view despite the theoretical advantages of terminal value monism (suffering is the only thing that motivates us “by itself”) over pluralism (there are many such things). Notably, ANU doesn’t require solving value incommensurability, because all other values can be instrumentally evaluated by their relationship to the suffering of sentient beings, using only one terminal value-grounded common currency for everything.
Therefore, it is a straw man argument that NUs don’t value life or positive states, because NUs value them instrumentally, which may translate into substantial practical efforts to protect them (compared even with someone who claims to be terminally motivated by them).
If the rationality and EA communities are looking for a unified theory of value, why are they not converging (more) on negative utilitarianism?
What have you read about it that has caused you to stop considering it, or to overlook it from the start?
Can you teach me how to see positive states as terminally (and not just instrumentally) valuable, if I currently don’t? (I still enjoy things, being closer to the extreme of hyperthymia than anhedonia. Am I platonically blind to the intrinsic aspect of positivity?)
And if someone wants to answer: What is the most extreme form of suffering that you’ve experienced and believe can be “outweighed” by positive experiences?