Eliezer_Yudkowsky comments on Decision Theory FAQ - Less Wrong

52 Post author: lukeprog 28 February 2013 02:15PM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (467)

You are viewing a single comment's thread. Show more comments above.

Comment author: Eliezer_Yudkowsky 13 March 2013 11:12:02PM 3 points [-]

And you'll never understand why we should all only make paperclips. (Where's Clippy when you need him?)

Comment author: davidpearce 14 March 2013 10:24:31AM 0 points [-]

Clippy has an off-the-scale AQ - he's a rule-following hypersystemetiser with a monomania for paperclips. But hypersocial sentients can have a runaway intelligence explosion too. And hypersocial sentients understand the mind of Mr Clippy better than Clippy understands the minds of sentients.

Comment author: TheOtherDave 14 March 2013 01:10:41PM 4 points [-]

And hypersocial sentients understand the mind of Mr Clippy better than Clippy understands the minds of sentients.

I'm confused by this claim.
Consider the following hypothetical scenario:

=======

I walk into a small village somewhere and find several dozen villagers fashioning paper clips by hand out of a spool of wire. Eventually I run into Clippy and have the following dialog.
"Why are those people making paper clips?" I ask.
"Because paper-clips are the most important thing ever!"
"No, I mean, what motivates them to make paper clips?"
"Oh! I talked them into it."
"Really? How did you do that?"
"Different strategies for different people. Mostly, I barter with them for advice on how to solve their personal problems. I'm pretty good at that; I'm the village's resident psychotherapist and life coach."
"Why not just build a paperclip-making machine?"
"I haven't a clue how to do that; I'm useless with machinery. Much easier to get humans to do what I want."
"Then how did you make the wire?"
"I didn't; I found a convenient stash of wire, and realized it could be used to manufacture paperclips! Oh joy!"

==========

It seems to me that Clippy in this example understands the minds of sentients pretty damned well, although it isn't capable of a runaway intelligence explosion. Are you suggesting that something like Clippy in this example is somehow not possible? Or that it is for some reason not relevant to the discussion? Or something else?

Comment author: whowhowho 14 March 2013 04:26:41PM *  2 points [-]

I think DP is saying that Clippy could not both understand suffering and cause suffering in the pursuit of clipping. The subsidiary arguments are:-

  • no entity can (fully) understand pain without empathising -- essentially, feeling it for itself.
  • no entity can feel pain without being strongly motivated by it, so an empathic clippy would be motivated against causing suffering.
  • And no, psychopaths therefore do not (fully) understand (others) suffering.
Comment author: TheOtherDave 14 March 2013 08:00:30PM 0 points [-]

I'm trying to figure out how you get from "hypersocial sentients understand the mind of Mr Clippy better than Clippy understands the minds of sentients" to "Mr Clippy could not both understand suffering and cause suffering in the pursuit of clipping" and I'm just at a loss for where to even start. They seem like utterly unrelated claims to me.

I also find the argument you quote here uncompelling, but that's largely beside the point; even if I found it compelling, I still wouldn't understand how it relates to what DP said or to the question I asked.

Comment author: whowhowho 14 March 2013 09:15:34PM -1 points [-]

"hypersocial sentients understand the mind of Mr Clippy better than Clippy understands the minds of sentients"

I (whowhowho) was not defending that claim.

"Mr Clippy could not both understand suffering and cause suffering in the pursuit of clipping"

To empathically understand suffering is to suffer along with someone who is suffering. Suffering has --or rather is -- negative value. An empath would not therefore cause suffering, all else being equal.

I'm just at a loss for where to even start.

Maybe don't restrict "understand" to "be able to model and predict".

Comment author: ArisKatsaris 15 March 2013 02:32:56PM 3 points [-]

Maybe don't restrict "understand" to "be able to model and predict".

If you want "rational" to include moral, then you're not actually disagreeing with LessWrong about rationality (the thing), but rather about "rationality" (the word).

Likewise if you want "understanding" to also include "empathic understanding" (suffering when other people suffer, taking joy when other people take joy), you're not actually disagreeing about understanding (the thing) with people who want to use the word to mean "modelling and predicting" you're disagreeing with them about "understanding" (the word).

Are all your disagreements purely linguistic ones? From the comments I've read of you so far, they seem to be so.

Comment author: davidpearce 15 March 2013 02:58:45PM *  2 points [-]

ArisKatsaris, it's possible to be a meta-ethical anti-realist and still endorse a much richer conception of what understanding entails than mere formal modeling and prediction. For example, if you want to understand what it's like to be a bat, then you want to know what the textures of echolocatory qualia are like. In fact, any cognitive agent that doesn't understand the character of echolocatory qualia-space does not understand bat-minds. More radically, some of us want to understand qualia-spaces that have not been recruited by natural selection to play any information-signalling role at all.

Comment author: whowhowho 15 March 2013 02:58:46PM 0 points [-]

I have argued that in practice, instrumental rationality cannot be maintained seprately from epistemic rationality, and that epistemic rationality could lead to moral objectivism, as many philosophers have argued. I don't think that those arguments are refuted by stipulatively defining "rationality" as "nothing to do with morality".

Comment author: TheOtherDave 14 March 2013 09:51:23PM 2 points [-]

I (whowhowho) was not defending that claim.

I quoted DP making that claim, said that claim confused me, and asked questions about what that claim meant. You replied by saying that you think DP is saying something which you then defended. I assumed, I think reasonably, that you meant to equate the thing I asked about with the thing you defended.

But, OK. If I throw out all of the pre-existing context and just look at your comment in isolation, I would certainly agree that Clippy is incapable of having the sort of understanding of suffering that requires one to experience the suffering of others (what you're calling a "full" understanding of suffering here) without preferring not to cause suffering, all else being equal.

Which is of course not to say that all else is necessarily equal, and in particular is not to say that Clippy would choose to spare itself suffering if it could purchase paperclips at the cost of its suffering, any more than a human would necessarily refrain from doing something valuable solely because doing so would cause them to suffer.

Comment author: davidpearce 14 March 2013 08:57:36PM 0 points [-]

Posthuman superintelligence may be incomprehensibly alien. But if we encountered an agent who wanted to maximise paperclips today, we wouldn't think, ""wow, how incomprehensibly alien", but, "aha, autism spectrum disorder". Of course, in the context of Clippy above, we're assuming a hypothetical axis of (un)clippiness whose (dis)valuable nature is supposedly orthogonal to the pleasure-pain axis. But what grounds have we for believing such a qualia-space could exist? Yes, we have strong reason to believe incomprehensibly alien qualia-spaces await discovery (cf. bats on psychedelics). But I haven't yet seen any convincing evidence there could be an alien qualia-space whose inherently (dis)valuable textures map on to the (dis)valuable textures of the pain-pleasure axis. Without hedonic tone, how can anything _matter _ at all?

Comment author: whowhowho 14 March 2013 09:17:37PM *  1 point [-]

But I haven't yet seen any convincing evidence there could be an alien qualia-space whose inherently (dis)valuable textures map on to the (dis)valuable textures of the pain-pleasure axis.

Meaning mapping the wrong way round, presumably.

Without hedonic tone, how can anything _matter _ at all?

Good question.

Comment author: TheOtherDave 14 March 2013 10:13:17PM 0 points [-]

if we encountered an agent who wanted to maximise paperclips today, we wouldn't think, ""wow, how incomprehensibly alien"

Agreed, as far as it goes. Hell, humans are demonstrably capable of encountering Eliza programs without thinking "wow, how incomprehensibly alien".

Mind you, we're mistaken: Eliza programs are incomprehensibly alien, we haven't the first clue what it feels like to be one, supposing it even feels like anything at all. But that doesn't stop us from thinking otherwise.

but, "aha, autism spectrum disorder".

Sure, that's one thing we might think instead. Agreed.

we're assuming a hypothetical axis of (un)clippiness whose (dis)valuable nature is supposedly orthogonal to the pleasure-pain axis. But what grounds have we for believing such a qualia-space could exist?

(shrug) I'm content to start off by saying that any "axis of (dis)value," whatever that is, which is capable of motivating behavior is "non-orthogonal," whatever that means in this context, to "the pleasure-pain axis," whatever that is.

Before going much further, though, I'd want some confidence that we were able to identify an observed system as being (or at least being reliably related to) an axis of (dis)value and able to determine, upon encountering such a thing, whether it (or the axis to which it was related) was orthogonal to the pleasure-pain axis or not.

I don't currently have any grounds for such confidence, and I doubt anyone else does either. If you think you do, I'd like to understand how you would go about making such determinations about an observed system.