I've been thinking about wireheading and the nature of my values. Many people here have defended the importance of external referents or complex desires. My problem is, I can't understand these claims at all.
To clarify, I mean wireheading in the strict "collapsing into orgasmium" sense. A successful implementation would identify all the reward circuitry and directly stimulate it, or do something equivalent. It would essentially be a vastly improved heroin. A good argument for either keeping complex values (e.g. by requiring at least a personal matrix) or external referents (e.g. by showing that a simulation can never suffice) would work for me.
Also, I use "reward" as short-hand for any enjoyable feeling, as "pleasure" tends to be used for a specific one of them, among bliss, excitement and so on, and "it's not about feeling X, but X and Y" is still wireheading after all.
I tried collecting all related arguments I could find. (Roughly sorted from weak to very weak, as I understand them, plus link to example instances. I also searched any literature/other sites I could think of, but didn't find other (not blatantly incoherent) arguments.)
- People do not always optimize their actions based on achieving rewards. (People also are horrible at making predictions and great at rationalizing their failures afterwards.)
- It is possible to enjoy doing something while wanting to stop or vice versa, do something without enjoying it while wanting to continue. (Seriously? I can't remember ever doing either. What makes you think that the action is thus valid, and you aren't just making mistaken predictions about rewards or are being exploited? Also, Mind Projection Fallacy.)
- A wireheaded "me" wouldn't be "me" anymore. (What's this "self" you're talking about? Why does it matter that it's preserved?)
- "I don't want it and that's that." (Why? What's this "wanting" you do? How do you know what you "want"? (see end of post))
- People, if given a hypothetical offer of being wireheaded, tend to refuse. (The exact result depends heavily on the exact question being asked. There are many biases at work here and we normally know better than to trust the majority intuition, so why should we trust it here?)
- Far-mode predictions tend to favor complex, external actions, while near-mode predictions are simpler, more hedonistic. Our true self is the far one, not the near one. (Why? The opposite is equally plausible. Or the falsehood of the near/far model in general.)
- If we imagine a wireheaded future, it feels like something is missing or like we won't really be happy. (Intuition pump.)
- It is not socially acceptable to embrace wireheading. (So what? Also, depends on the phrasing and society in question.)
(There have also been technical arguments against specific implementations of wireheading. I'm not concerned with those, as long as they don't show impossibility.)
Overall, none of this sounds remotely plausible to me. Most of it is outright question-begging or relies on intuition pumps that don't even work for me.
It confuses me that others might be convinced by arguments of this sort, so it seems likely that I have a fundamental misunderstanding or there are implicit assumptions I don't see. I fear that I have a large inferential gap here, so please be explicit and assume I'm a Martian. I genuinely feel like Gamma in A Much Better Life.
To me, all this talk about "valueing something" sounds like someone talking about "feeling the presence of the Holy Ghost". I don't mean this in a derogatory way, but the pattern "sense something funny, therefore some very specific and otherwise unsupported claim" matches. How do you know it's not just, you know, indigestion?
What is this "valuing"? How do you know that something is a "value", terminal or not? How do you know what it's about? How would you know if you were mistaken? What about unconscious hypocrisy or confabulation? Where do these "values" come from (i.e. what process creates them)? Overall, it sounds to me like people are confusing their feelings about (predicted) states of the world with caring about states directly.
To me, it seems like it's all about anticipating and achieving rewards (and avoiding punishments, but for the sake of the wireheading argument, it's equivalent). I make predicitions about what actions will trigger rewards (or instrumentally help me pursue those actions) and then engage in them. If my prediction was wrong, I drop the activity and try something else. If I "wanted" something, but getting it didn't trigger a rewarding feeling, I wouldn't take that as evidence that I "value" the activity for its own sake. I'd assume I suck at predicting or was ripped off.
Can someone give a reason why wireheading would be bad?
I believe that humans have natural psychological defenses against the lure of wireheading, because the appeal is something we navigate on a daily basis in our every day lives. In my case, I know I would really enjoy entertaining myself all the time (watching movies, eating good food, reading books) but eventually I would run out of money or feel guilty I'm not accomplishing anything.
Even if you tell people there will be no long-term consequences to wire-heading, they don't believe you. It's a matter of good character, actually, to be resistant to wanting to wirehead. For example, when people signal that they wouldn't wirehead because they prefer real interaction with external reality, there is reward (social and self-generated). (When I decide that I would 'never wirehead', I feel a sense of security and well-being.) I don't know why the intuitions don't work on you, perhaps you have a different set of background experiences so that you bypass the 'if-I-don't-stay-aware-now-I'll-lose-later' associations.
It seems to me that those who insist that they wouldn't wirehead, besides reaping social rewards for signaling a willingness to be altruistic, haven't fully taken to heart that values are not externally validated. If you have complex values x and y, they might as well be simple ones s and t. I think there are real (and healthy) physical/biological barriers to realizing this, so that even intellectuals won't become psychopaths.
But I agree with you it is how you phrase the question, and there are intuition pumps that pump the other way:
Futurists imagine a utopian future. But even if we achieved such, it wouldn't change the past. Why should the future time be so elevated in importance? No, any perfect future would be marred by the fact that for thousands of years in the past there was human suffering.
If the future can't be perfect due to the past, perhaps instead we could create a perfect sub-universe, where everything is perfect from beginning to end. Even if the outer universe can't be utopian, it can simulate utopias.
Then you might realize that by wireheading, you are simulating a utopian universe and thereby doing your part (one consciousness at a time) to create utopias. Then the only moral reason not to wirehead is if you think you can have enough influence to create more subjective happiness by not wireheadong than with your consciousness wireheaded alone.
Sometimes, my thoughts bend solopsistic (or at least simulation-based) and I wonder if the universe I'm in is already optimised. I'm skeptical because I hear there is suffering, but perhaps that is some necessary negativity to optimize my existence (part and parcel with having 'purpose'). I think I must actually believe this because stories of real suffering cause me intense disillusionment, as I don't expect it to be real and when I try to imagine it being real there is this strong resistance. I observe that many people seem to feel this way, if they're not in outright denial all the time about suffering being 'real'.
I was thinking earlier today -- I'll just throw this in -- that my intuition about values tends to be a little different than often described here because I feel that 1 consciousness with no suffering would be better than 10 consciousness with a mixed bag of experience. The only reason 10 consciousnesses might be better than 1 is because of their enjoyment of one another. So I guess I think quality is better than quantity, and also, I don't distinguish among consciousnesses. 10 moments of consciousness can be distributed over 10 people or 1 person, it doesn't matter. So if there really was a wire-heading machine, I might be convinced that not stepping in is equivalent to causing suffering to another person, equal to all the relative suffering I would encounter by not wireheading. However, I'm glad such choices are not available because -- for some reason, as I said I think it is biological -- wireheading just feels like a terrible negative choice, like entering a coffin. It feels isolated and not real.
If it's healthy to not be a psychopath, on what values do you base that? I think you're sneaking in a value judgment here that, if valid, would rule out wireheading.
(It might be evolutionary successful to not be a (full) psychopath, but that's a very different matter.)
I do find your overall thought process in your first few paragraphs plausible, but "anyone who disagrees with me is just not admitting that I'm right" sounds way too much like the kind of toxic reasoning I'm trying to avoid, so I'm fairly skeptical of it.