I don't think I've seen a model of identity that I like, which makes many thought experiments run into existential/identification issues. If you can answer "why is it acceptable (or not) to destroy a being and create a higher-experienced-utility being in it's place", then you have a basis for why it's OK for humans to change their own preferences by repetition or practice, but not OK for others to change them, or to change them by too much.
Sorry for the late response.
If people change their own preferences by repetition and practice, then they usually have a preference to do that. So it can be in their own best interests, for preferences they already have.
I could have a preference to change your preferences, and that could matter in the same way, but I don’t think I should say it's in your best interests (at least not for the thought experiment in this post). It could be in my best interests, or for whatever other goal I have (possibly altruistic).
In my view, identity preservation is vague and degreed, a matter of how much you inherit from your past "self", specifically how much of your memories and other dispositions.
Summary
Acknowledgements
Thanks to Lukas Gloor and Chi Nguyen for helpful feedback. Thanks to Teo Ajantaival, Magnus Vinding, Anthony DiGiovanni and Eleos Arete Citrini for helpful feedback on earlier related drafts. All errors are my own.
Manipulating desires and abandoning attachments
Let’s start with a thought experiment. Arneson (2006, pdf) wrote the following, although I substitute my own text in italics and square brackets to modify it slightly:
Assume further that Arneson loves Sam, his family and friends, philosophy and mountain biking, and would have continued to do so without the pill. He would have had a very satisfying, subjectively meaningful, personally fulfilling, pleasurable and happy life, with high levels of overall desire/preference satisfaction, even if he doesn’t take the pill. On all of these measures of subjective well-being, and any other general kind,[1] he will be very well off. These measures of subjective well-being are based on his attitudes, like his desires, his preferences, his pleasure and unpleasantness, his dispositions to enjoy things or find them unpleasant or aversive, his attachments, his judgements, his moral intuitions, and so on. By attitudes, I mean ways for things to seem to him to be good or bad, or better or worse, ways he cares about things. But he will be even better off on all general measures of subjective well-being, including overall desire/preference satisfaction, if he takes the pill, by assumption.
And all of this is based on his own accurate beliefs about his effects on and connection to the world; he’s not being fooled into believing he’s doing something he’s not or that any fake relationships he might have are real, like in Nozick’s experience machine (Nozick, 1974, pp.42–43, Nozick, 1989, pp.104–120, excerpts).
I’ll focus here on what’s best for Arneson.[2]
If subjective well-being is all that matters and higher lifetime subjective well-being is always better for someone, then for Arneson, it would be best for him to take the pill. Even if the frustration of his original desires, preferences or other attitudes is bad for him, we may just assume in the thought experiment that his gains with his new ones are greater than the losses on his original ones.[3]
And there’s nothing special about desiring only “[to know more about the world, so I will obsessively and happily consume scientific material (...)]”, which I’ve substituted. Arneson, in his thought experiment, used “casual sex, listening to sectarian religious sermons, mindless work, and TV watching”. It could be anything else, like becoming an entrepreneur, starting a new family in another country, making art (whether or not shared with others), meditating, counting blades of grass (Rawls, 1971, p. 432), watching paint dry or playing video games alone.
Rawls (1982, pdf, p.181) objected to the kind of verdict that Arneson should take the pill and the kinds of views delivering it, noting generally that they would make us bare persons:
It seems everything becomes replaceable. Can we really cherish our loved ones and our relationships with them if we would be willing to abandon them for whatever would bring us more well-being?[4]
However, Arneson could just grant that he is such a bare person and be willing to abandon his attitudes and attachments to increase his subjective well-being or subjective well-being in general. Arneson wouldn’t be honouring his specific attachments to his spouse, family and friends, philosophy or mountain biking, and perhaps he owes it to himself or to those attachments to honour them (Dorsey, 2019, full). But it’s also his life and these are his attitudes. Who am I to judge? Perhaps he has stronger attitudes towards his subjective well-being per se than towards his other specific attachments together.
On the other hand, if he does prefer not to take the pill and takes his particular attachments as important and worthy of honouring, more so than his potential gain in well-being per se, it seems to me that it wouldn’t be better for him and he shouldn’t take the pill. Again, it’s his life and these are his attitudes.
I expect most consequentialist welfarists, including most utilitarians, to go further in the opposite direction. On their views, not only would it be better for Arneson's welfare for him to take the pill, we should in principle force him take the pill even if he prefers not to, assuming this is the only other option, and ignoring effects on others and indirect reasons (which actually seem important to me in practice). Or, we should also give everyone else similar pills. Or, we could just kill everyone and replace them with things that generate more well-being.[5]
It doesn’t matter how much Arneson loves his spouse, his friends and family, his career or mountain biking, or how much he hates the prospect of giving them up or having his consent violated, as long as his gain in future well-being is large enough to outweigh the losses.
To me, changing someone’s attitudes this way against their wishes seems worse for them, if and because it’s worse according to the attitudes they already have or would otherwise have had.[6] These new attitudes should not be able to outweigh the losses in their current attitudes, just by ensuring the gain in subjective well-being is high enough. Such outweighing seems to violate the Platinum Rule, which would require us to treat others as they would have us treat them.[7] It may be the case that after we force Arneson to take the pill, we treated him as he (or his attitudes) would have had us treat him according to his new pill attitudes. However, while we are making him take the pill, we are definitely not treating him as he would have had us treat him at the time. And if we don’t force him to take the pill, we treat him as he would have us treat him all throughout; his pill attitudes never become real.
There is a substantial and artificial change in Arneson’s desires, preferences, perspective and attitudes towards things. His “utility function” changed.
Even if we had constraints against this sort of involuntary attitude manipulation — whether principled or for indirect or practical reasons — as long as we hold that, in the ideal, the involuntary attitude manipulation would be better for someone or better for the world, this betterness wouldn’t match their attitudes, the way they care about things, their subjective perspective. It only matches a new perspective, a new set of attitudes that’s been created.
In my view, this — weighing things to match and respect the attitudes of moral patients — is one of the most important things for moral views to get right, perhaps the most important. Otherwise, we’re projecting or imposing our own attitudes onto others. Instead, we should just listen to theirs.
And if Arneson doesn’t take the pill, he won’t have the pill-induced attitudes to which to listen.
Parfit (1984)[8], Yudkowsky (2007), Dorsey (2019, full)[9] also considered similar (voluntary) scenarios and thought experiments, and there are multiple related discussions of wireheading on LessWrong, especially Sotala (2010).
Responses and replies
People aren’t always right about what’s best for themselves
People aren’t always right in their judgements or beliefs about what’s best for themselves for various reasons:
Common to all three cases is that the individual neglects, discounts or otherwise misweighs attitudes they do, will or would have. So, it seems it can be right to disagree with them about what’s best for them, by better anticipating and weighing these attitudes they already have or will have.
Whether or not we should actually do things against their wishes for their benefit is a separate question I’ll discuss below. Still, if it would violate the Platinum rule to intervene on someone’s behalf against their judgement or allow it to happen like in one of these three cases, it could also violate it if we didn’t, because they would later have had us act differently, or a different discounted part of them, like any discounted feelings or desires, would have had us act differently.
We might think Arneson’s thought experiment is similar or even a special case of 2, and we are similarly justified in believing he’s wrong about what’s best for himself. If he takes the pill, this will be how he would have had us treat him, on all his later attitudes. He wouldn’t want to reverse the effects, by assumption, either.[10]
However, if Arneson doesn’t take the pill, he will not have such attitudes in favour of taking (or having taken) the pill, so it’s not a special case of 2. What he would neglect or discount — his attitudes in the counterfactual where he takes the pill — are things he does not and will not favour overall. Why should he care about these? Or why should we care about these on his behalf?[11]
Alternatively, there are no attitudes that will ever exist independently of the choice whether to take the pill that Arneson would neglect discount, or that we would neglect or discount on his behalf.[12]
Deontological constraints
One response would be deontological constraints or principled act-omission distinctions that prohibit you from manipulating the attitudes of others (or killing them) against their wishes and without their consent, even if it would increase their well-being or overall well-being. However, such merely personal constraints don’t give you reason to stop someone else from manipulating or killing others to increase well-being, which the Platinum rule could require of you.[13] But we could just have similar reasons to stop others, too. Then, we’re good right?
Either way, if manipulating someone’s attitudes against their wishes or consent would really be better for them if and because it increases their lifetime well-being, then it seems to me like you wouldn’t be stopping yourself from doing so for them. So what are these deontological constraints for?
I imagine some of those sympathetic to such deontological constraints have their own answers to this in turn, some perhaps grounded in indirect reasons related for welfare or individuals’ attitudes, like I discuss next. Some may not be grounded in such indirect reasons for welfare or individuals’ attitudes, and I’d object that this is projecting or imposing too much onto others, rather than listening to their attitudes.
Indirect reasons
Those sympathetic to welfarism of some kind can accept that it would be better for Arneson, given the assumptions, but deny that this thought experiment is informative of what we should ever really do. It might ignore more important indirect reasons for why it would tend to be bad for welfare:
I suspect these are indeed good reasons in practice.
Still, I also just deny that it really would be better for Arneson to take the pill, and specifically claim that it would be worse for him, or his attitudes. This requires a different conception of welfare. I will discuss this briefly in the last section, and more in another piece.
Incomparability
We could deny that Arneson’s original attitudes and attitudes after manipulation are comparable, because they are just too different. Or, generally, we could deny that different attitudes are comparable at all.
This would probably substantially undermine interpersonal comparisons, too, and make ethics much harder, although I think there is some good reason to believe in fairly widespread interpersonal incomparability, too (St. Jules, 2024b, St. Jules, 2024c). Arrow and Rawls rejected interpersonal comparisons (Arrow, 1977, p.225, Rawls, 1982, pdf, p.181).
However, this move doesn’t seem to imply that attitude manipulation is bad; it would just be incomparable to acts not manipulating attitudes. Arneson (2006, pdf) wrote:
Similarly, it seems we wouldn’t have reason to stop such attitude manipulation from happening — by accident or intentionally by someone else — in the first place.
Why should I care?
The thought experiment may seem too removed from practicality in the near term or ever. Brains are hard to manipulate reliably like this. Why should you care about such a thought experiment, even if you agree that taking the pill wouldn’t be better for Arneson?
Something similar seems reasonably likely to be possible in the future with more advanced technology. At least, it seems quite likely if and when we can reprogram artificially conscious beings — including digital people and mind uploads, as well as conscious minds not specifically much like humans’ — to change their attitudes.
Perhaps more importantly, whether or not we can ever carry out something like the thought experiment, if we admit that such attitude manipulation can be wrong to or worse for the affected individual even if it increases their subjective well-being, then this should be reflected in our moral views. In particular, I believe it supports certain kinds of person-affecting views, as I will argue briefly in the next section and further in another piece. Such moral views could have important implications for cause prioritization, especially reducing the priority given to extinction risk reduction.
Avoiding manipulation with preference-affecting views
What could work, then? A principled response according to which (especially involuntary) attitude manipulation is directly bad for you would need to
This sounds like what a preference-affecting view can accomplish. Preference-affecting views are in favor of making preferences (or attitudes) satisfied, but neutral about making satisfied preferences (Barry, 1989, Bykvist, 1996 and Dietz, 2023).
Compare to person-affecting views. Person-affecting views are “in favor of making people happy, but neutral about making happy people” (Narveson, 1973, p.80). The neutrality towards “making happy people” means denying that we have any reason to create people for their own sake, although there may be indirect reasons. There are different ways of understanding “making people happy” that correspond to different person-affecting views. Should we only care about people who already exist (presentism)?[14] Should we only care about people who would exist regardless of our choice (necessitarianism)?[15] Should we care about how the lives of future people go in other ways besides ensuring more come to exist at all (various other views)?[16]
On a person-affecting view,[17] if killing someone is bad for them, say because it deprives them of potential future goods or frustrates their preferences, creating other happy people wouldn’t make up for it, all else equal. So, person-affecting views can take more principled stances against killing and replacing, although not all person-affecting views actually achieve this (St. Jules, 2024a). Mind uploading and cryonics would also make more sense as impartial pursuits on person-affecting (and preference-affecting) views.
Preference-affecting views are basically person-affecting views, but applied to individual preferences (or attitudes) instead of just to whole individuals or whole individual welfare. We can generate a preference-affecting view from a specific person-affecting view by applying the rules of the person-affecting view to attitudes instead of to whole individuals or their welfare. So, we could care only for attitudes that currently exist (presentism), only for those that would exist regardless of our choice (necessitarianism), or also for contingent attitudes in other ways besides ensuring more come to exist at all.
Preference-affecting views also are in fact person-affecting views. As such, causes are likely to be prioritized like on similar person-affecting views. Perhaps most notably, I expect extinction risk reduction to receive less priority on these views. If (human or total animal) extinction is bad at all, it would not be because of the positive value in the foregone future contingent lives, but depending on the particular view, if and because it’s bad for current individuals,[18] leads to more overall bad lives (on asymmetric views), is bad for other future individuals (e.g. other animals, aliens or artificial conscious beings) or replaces better off contingent individuals with worse off ones. But extinction could be good for opposite reasons, too.
Another possibility would be to recognize both preference-affecting reasons and totalist reasons together. This would raise the bar for but still allow attitude manipulation and replacement. So, for example, it could still sometimes be better for you to abandon your loved ones even if you prefer not to, but the net gain in subjective well-being would have to pass some threshold; not just any positive net gain is enough. Similarly, it could still be better if everyone was killed and replaced with better off beings, but the gain in well-being would have to be large enough. And similarly, to prioritize ensuring future moral patients exist for their own sake or the sake of their well-being would require the gain in well-being to be large enough, and not just larger than the opportunity costs for existing individuals.
My own view is that such a hybrid approach still does too much projecting or imposing onto others or the world. We should just listen.
In my next piece, I will further motivate, sketch and defend fully preference-affecting views.
As opposed to narrower ones with types of objects of his attitudes specified, e.g. satisfaction with his relationships with others.
If Arneson taking the pill would be bad for Arneson’s spouse, family, friends or others in the world, we could:
1. consider those impacts separately,
2. assume his gains in well-being are greater than the losses to others, or
3. give the others similar pills, giving up any attachments to Arneson and increasing their subjective well-being, too.
This rules out solutions based just on global preferences, i.e. an individual’s preferences about their own life as a whole, or whole outcomes, the way things go (Parfit, 1984), and hidden desires, or implicit preferences, e.g. to not have one’s preferences changed in certain ways (Yu, 2022).
Again, assuming they aren’t made worse off, or they take similar pills, or that you only care about them through your own subjective well-being, if you’re willing to manipulate it.
The apparent betterness of attitude manipulation follows from letting newly created attitudes offset or outweigh the frustration already existing or otherwise necessary attitudes. The original attitudes become replaceable. This has the same structure as the problem of replacement or replaceability in population ethics (Singer, 1979a, Singer, 1979b, Jamieson, 1984, Knutsson, 2021, St. Jules, 2024a), in which (potentially very well off) individuals are killed and replaced with (on aggregate) better off individuals. In our case here, replacement is applied between attitudes within an individual instead of between whole individuals. Furthermore, killing and replacing just seems like a more extreme form of attitude manipulation, where the individual’s personal identity — the characteristics that make someone the same person over time and across possible worlds (Olsen, 2023, Shoemaker, 2019 and/or Mackie & Jago, 2022) — is not at all preserved, even partially.
If on the attitudes they will have without taking the pill, taking the pill would have been better, say because without the pill, they will suffer a lot more or wish they had never been born, or wish they had taken the pill, then it seems plausible to me that it would be better to take the pill. I will expand on such a view in my next piece.
The Platinum Rule can be seen as a variant of the Golden Rule, which would require us to treat others as we would have them treat us. Or, it can be seen as an application of the Golden Rule, because we would have others treat us in accordance with our attitudes, so we should treat others in accordance with their attitudes.
Parfit (1984) wrote:
Dorsey (2019, full) wrote:
This is an example of the reversal test (Bostrom & Ord, 2006).
However, we could also imagine another pill to reverse the effects of the first, and him reuniting with his spouse, friends and family and taking up philosophy and mountain biking again. If he does take the second pill after the first, he could judge, back with his original attitudes and life and knowledge of what happened, that it was a mistake to take the first pill.
This motivates an actualist preference-affecting (or attitude-affecting) view. I will flesh out such a view and its implications in my next piece.
In line with a necessitarian preference-affecting view.
Others wouldn’t necessarily particularly care whether it’s you or someone else manipulating or killing them or whether you stop someone else or stop yourself, so it isn’t clear why you wouldn’t have similarly strong reasons to stop someone else from manipulating or killing others to increase wellbeing, especially if those reasons are really just about others. A personal deontological constraint, if not extended to stopping others from violating it, seems partly self-regarding.
No, in my view.
No, in my view.
Yes, in my view.
Assuming we can say that someone is still (roughly) the same person in the future and across possible worlds, as opposed to there being no fact of the matter or that we are constantly changing identities. For standard references on personal identity, see Olsen, 2023, Shoemaker, 2019 and/or Mackie & Jago, 2022.
E.g. Finneron-Burns, 2017. Depending on advances in technology like life extension and mind uploading, and how we count the future welfare of currently existing people, currently existing people could live extremely long lives and so be made far better (or worse) off (Gustafsson & Kosonen, 2019, Shulman, 2019).