You're looking at Less Wrong's discussion board. This includes all posts, including those that haven't been promoted to the front page yet. For more information, see About Less Wrong.

AlephNeil comments on Lifeism, Anti-Deathism, and Some Other Terminal-Values Rambling - Less Wrong Discussion

4 Post author: Pavitra 07 March 2011 04:35AM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (87)

You are viewing a single comment's thread.

Comment author: AlephNeil 07 March 2011 05:32:37PM *  3 points [-]

I don't think it's possible to give answers to all ethical dilemmas in such a way as to be consistent and reasonable across the board, but here my intuition is that if a mind only lasts 15 minutes, and it has no influence on the outside world and leaves no 'thought children' (e.g. doodles, poems, theorems) behind after its death, then whether it experiences contentment or agony has no moral value whatsoever. Its contentment, its agony, its creation and its destruction are all utterly insignificant and devoid of ethical weight.

To create a mind purely to torture it for 15 minutes is something only an evil person would want to do (just as only an evil person would watch videos of torture for fun) but as an act, it's a mere 'symptom' of the fact that all is not well in the universe.

(However, if you were to ask "what if the person lasted 30 minutes? A week? A year? etc." then at some point I'd have to change my answer, and it might be difficult to reconcile both answers. But again, I don't believe that the 'sheaf' of human moral intuitions has a 'global section'.)

the net utility of the existence of a group of entities-whose-existence-constitutes-utility is equal to the maximum of the individual utilities

Hmm. There might be a good insight lurking around there, but I'd want to argue that (a) such entities may include 'pieces of knowledge', 'trains of thought', 'works of art', 'great cities' etc rather than just 'people'. And (b), the 'utilities' (clearer to just say 'values') of these things might be partially rather than linearly ordered, so that the 'maximum' becomes a 'join', which may not be attained by any of them individually. (Is the best city better or worse than the best symphony, and are they better or worse than Wiles' proof of Fermat's Last Theorem, and are they better or worse than a giraffe?)

Comment author: DanielVarga 07 March 2011 08:24:45PM 1 point [-]

I agree fully with your first two paragraphs. I would not change my answer regardless of the amount of time the causally disconnected person lasts. Biting this bullet leads to some quite extreme conclusions, basically admitting that current human values can not be consistently transferred to a future with uploads, self-modification and such. (Meaning, Eliezer's whole research program is futile.) I am not happy about these conclusions, but they do not change my respect for human values, regardless of my opinion about their fundamental inconsistencies.

I believe even AlephNeil's position is quite extreme among LWers, and mine is definitely fringe. So if someone here agrees with either of us, I am very interested in that information.

Comment author: endoself 08 March 2011 04:44:49AM 1 point [-]

Biting this bullet leads to some quite extreme conclusions, basically admitting that current human values can not be consistently transferred to a future with uploads, self-modification and such. (Meaning, Eliezer's whole research program is futile.)

Couldn't an AI prevent us from ever achieving uploads or self-modification? Wouldn't this be a good thing for humanity if human values could not survive in a future with those things?

Comment author: DanielVarga 08 March 2011 10:56:59PM *  1 point [-]

Yes, this is a possible end point of my line of reasoning: we either have to become luddites, or build a FAI that prevents us from uploading. These are both very repulsive conclusions for me. (Even if I don't consider the fact that I am not confident enough in my judgement to justify such extreme solutions by it.) I, personally, would rather accept that much of my values will not survive.

My value system works okay right now, at least when I don't have to solve trolley problems. In any given world with uploading and self-modification, my value system would necessarily fail. In such a world, my current self would not feel at home. My visit there would be a series of unbelievably nasty trolley problems, a big reductio ad absurdum of my values. Luckily, it is not me who has to feel at home there, but the inhabitants of that world. (*)

(*) Even the word "inhabitants" is misleading, because I don't think personal identity has much of a role in a world where it is possible to merge minds. Not to talk about the word "feel", which, from the perspective of a substrate-independent self-modifying mind refers to a particular suboptimal self-reflection mechanism. Which, to clear up a possible misunderstanding in advance, does not mean that this substrate-independent mind can not possibly see positive feelings as terminal value. But I am already quite off-topic here.

Comment author: endoself 08 March 2011 11:22:31PM 0 points [-]

I, personally, would rather accept that much of my values will not survive.

If there is something that you care about more than your values, they are not really your values.

I think we should just get on with FAI. If it realizes that uploads are okay according to our values it will allow uploads and if uploads are bad it will forbid them (maybe not entirely forbid; there could easily be something even worse). This is one of the questions that can completely be left until after we have FAI because whatever it does will, by definition, be in accordance with our values.

Comment author: Pavitra 09 March 2011 12:56:36AM 0 points [-]

I, personally, would rather accept that much of my values will not survive.

If there is something that you care about more than your values, they are not really your values.

You seem to conflate "I will care about X" with "X will occur". This breaks down in, for example, any case where precommitment is useful.

Comment author: DanielVarga 09 March 2011 12:42:48AM 0 points [-]

If there is something that you care about more than your values, they are not really your values.

You seem to rely on a hidden assumption here: that I am equally confident in all my values.

I don't think my values are consistent. Having more powerful deductive reasoning, and constant access to extreme corner cases would obviously change my value system. I also anticipate that my values would not be changed equally. Some of them would survive the encounter with extreme corner cases, some would not. Right now I don't have to constantly deal with perfect clones and merging minds, so I am fine with my values as they are. But even now, I have a quite good intuition about which of them would not survive the future shock. That's why I can talk without contradiction about accepting to lose those.

In CEV jargon: my expectation is that the extrapolation of my value system might not be recognizable to me as my value system. Wei_Dai voiced some related concerns with CEV here. It is worth looking at the first link in his comment.

Comment author: endoself 09 March 2011 03:03:05AM 0 points [-]

Oh, I see. I appear to have initially missed the phrase `much of my values'.

I am wary of referring to my current inconsistent values rather than their reflective equilibrium as `my values' because of the principle of explosion, but I am unsure of how to resolve this into my current self even having values.

Comment author: DanielVarga 09 March 2011 11:16:57AM *  1 point [-]

It seems our positions can be summed up like this: You are wary of referring to your current values rather than their reflective equilibrium as 'your values', because your current values are inconsistent. I am wary of referring to the reflective equilibrium rather than my current values as 'my values', because I expect the transition to reflective equilibrium to be a very aggressive operation. (One could say that I embrace my ignorance.)

My concern is that the reflective equilibrium is far from my current position in the dynamical system of values. Meanwhile, Marcello and Wei Dai are concerned that the dynamical system is chaotic and has multiple reflective equilibria.

Comment author: endoself 09 March 2011 08:20:56PM *  0 points [-]

I don't worry about the aggressiveness of the transition because, if my current values are inconsistent, they can be made to say that this transition is both good and bad. I share the concern about multiple reflective equilibrium. What does it mean to judge something as an irrational cishuman if two reflective equilibria would disagree on what is desirable?

Comment author: TheOtherDave 09 March 2011 04:24:20PM 0 points [-]

I expect the transition to reflective equilibrium to be a very aggressive operation.

Upvoted purely for the tasty, tasty understatement here.

I should get that put on a button.

Comment author: Pavitra 09 March 2011 03:22:21AM 0 points [-]

I like to think of my "true values" as (initially) unknown, and my moral intuitions as evidence of, and approximations to, those true values. I can then work on improving the error margins, confidence intervals, and so forth.

Comment author: endoself 09 March 2011 03:36:50AM 0 points [-]

So do I, but I worry that they are not uniquely defined by the evidence. I may eventually be moved to unique values by irrational arguments, but if those values are different from my current true values than I will have lost something and if I don't have any true values than my search for values will have been pointless, though my future self will be okay with that.

Comment author: Pavitra 08 March 2011 04:34:12AM 0 points [-]

Your point about partial ordering is very powerfully appealing.

However, I feel that any increase in utility from mere accumulation tends strongly to be completely overridden by increase in utility from increasing the quality of the best thing you have, such as by synthesizing a symphony and a theorem together into some deeper, polymathic insight. There might be edge cases where a large increase in quantity outweighs a small increase in quantity, but I haven't thought of any yet.

(Incidentally, I just noticed that I've been using terms incorrectly and I'm actually a consequentialist rather than a utilitarian. What should I be saying instead of "utility" to mean that-thing-I-want-to-maximize?)