Wei_Dai comments on Why I Moved from AI to Neuroscience, or: Uploading Worms - Less Wrong

43 Post author: davidad 13 April 2012 07:10AM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (56)

You are viewing a single comment's thread. Show more comments above.

Comment author: davidad 13 April 2012 10:57:44PM 6 points [-]

My answers are indeed "the latter" and "yes". There are a couple ways I can justify this.

The first way is just to assert that from a standard utilitarian perspective, over the long term, technological progress is a fairly good indicator for lack of suffering (e.g. Europe vs. Africa). [Although arguments have been made that happiness has gone down since 1950 while technology has gone up, I see the latter 20th century as a bit of a "dark age" analogous to the fall of antiquity (we forgot how to get to the moon!) which will be reversed in due time.]

The second is that I challenge you to define "pleasure," "happiness," or "lack of suffering." You may challenge me to define "technological progress," but I can just point you to sophistication or integrated information as reasonable proxies. As vague as notions of "progress" and "complexity" are, I assert that they are decidedly less vague than notions of "pleasure" and "suffering". To support this claim, note that sophistication and integrated information can be defined and evaluated without a normative partition of the universe into a discrete set of entities, whereas pleasure and suffering cannot. So the pleasure metric leads to lots of weird paradoxes. Finally, self-modifying superintelligences must necessarily develop a fundamentally different concept of pleasure than we do (otherwise they just wirehead), so the pleasure metric probably cannot be straightforwardly applied to their situation anyway.

Comment author: Wei_Dai 14 April 2012 12:40:33AM *  13 points [-]

The first way is just to assert that from a standard utilitarian perspective, over the long term, technological progress is a fairly good indicator for lack of suffering (e.g. Europe vs. Africa).

What about hunter-gatherers vs farmers? And a universe devoid of both life and technology would have even less suffering than either.

The second is that I challenge you to define "pleasure," "happiness," or "lack of suffering."

Can you explain why you're giving me this challenge? Because I don't understand, if I couldn't define them except vaguely, how does it strengthen your case that we should care about technology and not these values. Suppose I told you that I want to maximize the smoothness of the universe, because that's even easier to define than "technology"? Wouldn't you think that's absurd?

Edit: Also, could you clarify whether you value technology as an end in itself, or just as a proxy for for your real values which perhaps you can't easily define but might be something like "life being good"?

Comment author: XiXiDu 14 April 2012 10:10:35AM 2 points [-]

The second is that I challenge you to define "pleasure," "happiness," or "lack of suffering."

Can you explain why you're giving me this challenge? Because I don't understand, if I couldn't define them except vaguely, how does it strengthen your case that we should care about technology and not these values.

As far as I understand him, he is saying that technological progress can be quantified. While all your ideas of how to rate world states can either not be quantified, and therefore can't be rated, or run into problems and contradictions.

He further seems to believe that technological progress leads to "complexity" which leads to other kinds of values. Even if they are completely alien to us humans and our values, they will still be intrinsically valuable.

His view of a universe where an "unfriendly" AI takes over is a universe where there will be a society of paperclip maximizer's and their offspring. Those AI's will not only diverge from maximizing paperclips, and evolve complex values, but also pursue various instrumental goals, as exploration will never cease. And pursuing those goals will satisfy their own concept of pleasure.

And he believes that having such a culture of paperclip maximizer's having fun while pursuing their goals isn't less valuable than having our current volition being extrapolated, which might end up being similarly alien to our current values.

In other words, there is one thing that we can rate and that is complexity. If we can increase it then we should do so. Never mind the outcome, it will be good.

Correct me if I misinterpreted anything.

Comment author: davidad 14 April 2012 12:22:25PM 1 point [-]

I couldn't have said it better myself.

Comment author: Wei_Dai 14 April 2012 02:46:50PM 7 points [-]

Would you change your mind if I could give a precise definition of, say, "suffering", and showed you two paths to the future that end up with similar levels of technology but different amounts of suffering? I'll assume the answer is yes, because otherwise why did you give me that challenge.

What if I said that I don't know how to define it now, but I think if you made me a bit (or a lot) smarter and gave me a few decades of subjective time to work on the problem, I could probably give you such a definition and tell you how to achieve the "less suffering, same tech" outcome? Would you be willing to give me that chance (assuming it was in your power to do so)? Or are you pretty sure that "suffering" is not just hard to define, but actually impossible, and/or that it's impossible to reduce suffering to any significant extent below the default outcome, while keeping technology at the same level? If you are pretty sure about this, are you equally sure about every other value that I could cite instead of suffering?

Comment author: XiXiDu 14 April 2012 03:55:51PM 2 points [-]

Or are you pretty sure that "suffering" is not just hard to define, but actually impossible, and/or that it's impossible to reduce suffering to any significant extent below the default outcome, while keeping technology at the same level?

Masochist: Please hurt me!

Sadist: No.

If you are pretty sure about this, are you equally sure about every other value that I could cite instead of suffering?

Not sure, but it might be impossible.

What if I said that I don't know how to define it now, but I think if you made me a bit (or a lot) smarter...

If you were to uplift a chimpanzee onto the human level and told it to figure out how to reduce suffering for chimpanzees, it would probably come up with ideas like democracy, health insurance and supermarkets. Problem is that chimpanzees wouldn't appreciate those ideas...

Comment author: Wei_Dai 14 April 2012 09:52:41PM 4 points [-]

XiXiDu, I'm aware that I'm hardly making a watertight case that I can definitely do better than davidad's plan (from the perspective of his current apparent values). I'm merely trying to introduce some doubt. (Note how Eliezer used to be a technophile like David, and said things like "But if it comes down to Us or Them, I'm with Them.", but then changed his mind.)