Will_Newsome comments on Our Phyg Is Not Exclusive Enough - Less Wrong

25 [deleted] 14 April 2012 09:08PM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (513)

You are viewing a single comment's thread. Show more comments above.

Comment author: Will_Newsome 15 April 2012 09:23:14AM 6 points [-]
Comment author: wedrifid 15 April 2012 09:47:40AM 4 points [-]

From my favorite post and comments section on Less Wrong thus far:

Take metaethics, a solved problem: what are the odds that someone who still thought metaethics was a Deep Mystery could write an AI algorithm that could come up with a correct metaethics?

Yes, it looks like Eliezer is mistaken there (or speaking hyperbole).

I agree with:

what are the odds that someone who still thought metaethics was a Deep Mystery could write an AI algorithm that could come up with a correct metaethics?

... but would weaken the claim drastically to "Take metaethics, a clearly reducible problem with many technical details to be ironed out". I suspect you would disagree with even that, given that you advocate meta-ethical sentiments that I would negatively label "Deeply Mysterious". This places me approximately equidistant from your respective positions.

Comment author: Will_Newsome 15 April 2012 09:57:40AM 4 points [-]

I only weakly advocate certain (not formally justified) ideas about meta-ethics, and remain deeply confused about certain meta-ethical questions that I wouldn't characterize as mere technical details. One simple example: Eliezer equates reflective consistency (a la CEV) with alignment with the big blob of computation he calls "right"; I still don't know what argument, technical or non-technical, could justify such an intuition, and I don't know how Eliezer would make tradeoffs if the two did in fact have different referents. This strikes me as a significant problem in itself, and there are many more problems like it.

(Mildly inebriated, apologies for errors.)

Comment author: gjm 15 April 2012 11:19:56PM 4 points [-]

Are you sure Eliezer does equate reflective consistency with alignment with what-he-calls-"right"? Because my recollection is that he doesn't claim either (1) that a reflectively consistent alien mind need have values at all like what he calls right, or (2) that any individual human being, if made reflectively consistent, would necessarily end up with values much like what he calls right.

(Unless I'm awfully confused, denial of (1) is an important element in his thinking.)

I think he is defining "right" to mean something along the lines of "in line with the CEV of present-day humanity". Maybe that's a sensible way to use the word, maybe not (for what it's worth, I incline towards "not") but it isn't the same thing as identifying "right" with "reflectively consistent", and it doesn't lead to a risk of confusion if the two turn out to have different referents (because they can't).

Comment author: thomblake 16 April 2012 09:26:01PM 1 point [-]

Eliezer equates reflective consistency (a la CEV) with alignment with the big blob of computation he calls "right"

He most certainly does not.

Comment author: steven0461 17 April 2012 02:07:34AM 2 points [-]

Relevant quote from Morality as Fixed Computation:

But the key notion is the idea that what we name by 'right' is a fixed question, or perhaps a fixed framework. We can encounter moral arguments that modify our terminal values, and even encounter moral arguments that modify what we count as a moral argument; nonetheless, it all grows out of a particular starting point. We do not experience ourselves as embodying the question "What will I decide to do?" which would be a Type 2 calculator; anything we decided would thereby become right. We experience ourselves as asking the embodied question: "What will save my friends, and my people, from getting hurt? How can we all have more fun? ..." where the "..." is around a thousand other things.

So 'I should X' does not mean that I would attempt to X were I fully informed.

Comment author: thomblake 17 April 2012 12:24:18PM 0 points [-]

Thanks - I hope you're providing that as evidence for my point.

Comment author: steven0461 17 April 2012 08:10:41PM *  0 points [-]

Sort of. It certainly means he doesn't define morality as extrapolated volition. (But maybe "equate" meant something looser than that?)

Comment author: Will_Newsome 20 April 2012 05:19:44AM 1 point [-]

Aghhhh this is so confusing. Now I'm left thinking both you and Wei Dai have furnished quotes supporting my position, User:thomblake has interpreted your quote as supporting his position, and neither User:thomblake nor User:gjm have replied to Wei Dai's quote so I don't know if they'd interpret it as evidence of their position too! I guess I'll just assume I'm wrong in the meantime.

Comment author: Will_Newsome 17 April 2012 01:21:06AM 0 points [-]

Now two people have said the exact opposite things both of which disagree with me. :( Now I don't know how to update. I plan on re-reading the relevant stuff anyway.

Comment author: gjm 17 April 2012 01:37:55AM 0 points [-]

If you mean me and thomblake, I don't see how we're saying exact opposite things, or even slightly opposite things. We do both disagree with you, though.

Comment author: Will_Newsome 17 April 2012 01:47:53AM 2 points [-]

I guess I can interpret User:thomblake two ways, but apparently my preferred way isn't correct. Let me rephrase what you said from memory. It was like, "right is defined as the output of something like CEV, but that doesn't mean that individuals won't upon reflection differ substantially". User:thomblake seemed to be saying "Eliezer doesn't try to equate those two or define one as the other", not "Eliezer defines right as CEV, he doesn't equate it with CEV". But you think User:thomblake intended the latter? Also, have I fairly characterized your position?

Comment author: thomblake 17 April 2012 12:14:34PM *  1 point [-]

Eliezer defines right as CEV, he doesn't equate it with CEV

I definitely meant the latter, and I might be persuaded of the former.

Though "define" still seems like the wrong word. More like, " 'right' is defined as *point at big blob of poetry*, and I expect it will be correctly found via the process of CEV." - but that's still off-the-cuff.

Comment author: Will_Newsome 17 April 2012 12:26:47PM 1 point [-]

Thanks much; I'll keep your opinion in mind while re-reading the meta-ethics sequence/CEV/CFAI. I might be being unduly uncharitable to Eliezer as a reaction to noticing that I was unduly (objectively-unjustifiably) trusting him. (This would have been a year or two ago.) (I notice that many people seem to unjustifiably disparage Eliezer's ideas, but then again I notice that many people seem to unjustifiably anti-disparage (praise, re-confirm, spread) Eliezer's ideas;—so I might be biased.)

(Really freaking drunk, apologies for errors, e.g. poltiically unmotivated adulation/anti-adulation, or excessive self-divulgation. (E.g., I suspect "divulgation" isn't a word.))

Comment author: thomblake 17 April 2012 12:29:23PM 1 point [-]

I suspect "divulgation" isn't a word.

Not to worry, it means "The act of divulging" or else "public awareness of science" (oddly).

Comment author: gjm 17 April 2012 10:08:33AM 1 point [-]

I don't know whether thomblake intended the latter, but he certainly didn't say the former. I think you said "Eliezer said A and B", thomblake said "No he didn't", and you are now saying he meant "Eliezer said neither A nor B". I suggest that he said, or at least implied, something rather like A, and would fiercely repudiate B.

Comment author: wedrifid 15 April 2012 10:09:27AM 1 point [-]

One simple example: Eliezer equates reflective consistency (a la CEV) with alignment with the big blob of computation he calls "right"; I still don't know what argument, technical or non-technical, could justify such an intuition, and I don't know how Eliezer would make tradeoffs if the two did in fact have different referents.

If I understand you correctly then this particular example I don't think I have a problem with, at least not when I assume the kind of disclaimers and limitations of scope that I would include if I were to attempt to formally specify such a thing.

This strikes me as a significant problem in itself, and there are many more problems like it.

I suspect I agree with some of your objections to various degrees.

Comment author: Wei_Dai 15 April 2012 10:42:13PM 7 points [-]

Part of my concern about Eliezer trying to build FAI also stems from his treatment of metaethics. Here's a caricature of how his solution looks to me:

Alice: Hey, what is the value of X?

Bob: Hmm, I don't know. Actually I'm not even sure what it means to answer that question. What's the definition of X?

Alice: I don't know how to define it either.

Bob: Ok... I don't know how to answer your question, but what if we simulate a bunch of really smart people and ask them what the value of X is?

Alice: Great idea! But what about the definition of X? I feel like we ought to be able to at least answer that now...

Bob: Oh that's easy. Let's just define it as the output of that computation I just mentioned.

Comment author: amit 16 April 2012 12:08:27AM *  1 point [-]

I thought the upshot of Eliezer's metaethics sequence was just that "right" is a fixed abstract computation, not that it's (the output of) some particular computation that involves simulating really smart people. CEV is not even mentioned in the sequence (EDIT: whoops it is.).

(Indeed just saying that it's a fixed abstract computation is at the right level of abstraction to qualify as metaethics; saying that it's some particular computation would be more like just plain ethics. The upshot does feel kind of underwhelming and obvious. This might be because I just don't remember how confusing the issue looked before I read those posts. It could also mean that Eliezer claiming that metaethics is a solved problem is not as questionable as it might seem. And it could also mean that metaethics being solved doesn't consitute as massive progress as it might seem.)

Comment author: Wei_Dai 17 April 2012 06:56:00AM *  7 points [-]

The upshot does feel kind of underwhelming and obvious. This might be because I just don't remember how confusing the issue looked before I read those posts.

BTW, I've had numerous "wow" moments with philosophical insights, some of which made me spend years considering their implications. For example:

  • Bayesian interpretation of probability
  • AI / intelligence explosion
  • Tegmark's mathematical universe
  • anthropic principle / anthropic reasoning
  • free will as the ability to decide logical facts

I expect that a correct solution to metaethics would produce a similar "wow" reaction. That is, it would be obvious in retrospect, but in an overwhelming instead of underwhelming way.

Comment author: stoat 17 April 2012 02:26:28PM 1 point [-]

Is the insight about free will and logical facts part of the sequences? or is it something you or others discuss in a post somewhere? I'd like to learn about it, but my searches failed.

Comment author: Wei_Dai 17 April 2012 09:18:02PM *  2 points [-]

I never wrote a post on it specifically, but it's sort of implicit in my UDT post (see also this comment). Eliezer also has a free will sequence which is somewhat similar/related but I'm not sure if he would agree with my formulation.

Comment author: XiXiDu 18 April 2012 08:41:44AM -1 points [-]

"What is it that you're deciding when you make a decision?"

What is "you"? And what is "deciding"? Personally I haven't been able to come to any redefinition of free will that makes more sense than this one.

I haven't read the free will sequence. And I haven't read up on decision theory because I wasn't sure if my math education is good enough yet. But I doubt that if I was going to read it I would learn that you can salvage the notion of "deciding" from causality and logical facts. The best you can do is look at an agent and treat it is as a transformation. But then you'd still be left with the problem of identity.

Comment author: Will_Newsome 17 April 2012 07:53:29AM *  -2 points [-]

(Agreed; I also think meta-ethics and ethics are tied into each other in a way that would require that a solution to meta-ethics would at least theoretically solve any ethical problems. Given that I can think of hundreds or thousands of object level ethical problems, and given that I don't think my inability to answer at least some of them is purely due to boundedness, fallibility, self-delusion, or ignorance as such, I don't think I have a solution to meta-ethics. (But I would characterize my belief in God as at least a belief that meta-ethics and ethical problems do at least have some unique (meta-level) solution. This might be optimistic bias, though.))

Wei Dai, have you read the Sermon on the Mount, particularly with superintelligences, Tegmark, (epistemic or moral) credit assignment, and decision theory in mind? If not I suggest it, if only for spiritual benefits. (I suggest the Douay-Rheims translation, but that might be due to a bias towards Catholics as opposed to Protestants.)

(Pretty damn drunk for the third day in a row, apologies for errors.)

Comment author: khafra 18 April 2012 05:55:09AM 0 points [-]

Are you planning on starting a rationalist's drinking club? A byob lesswrong meetup with one sober note-taker? You usually do things purposefully, even if they're unusual purposes, so consistent drunkenness seems uncharacteristic unless it's part of a plan.

Comment author: wedrifid 18 April 2012 06:10:49AM 1 point [-]

Are you planning on starting a rationalist's drinking club?

Will_Newsome isn't a rationalist. (He has described himself as a 'post-rationalist', which seems as good a term as any.)

Comment author: Will_Newsome 17 April 2012 09:06:18AM -1 points [-]

In vino veritas et sanitas!

Comment author: Wei_Dai 16 April 2012 07:48:57AM *  1 point [-]

CEV is not even mentioned in the sequence.

It's mentioned here:

So here's my metaethics:

[...]

For a human this is a much huger blob of a computation that looks like, "Did everyone survive? How many people are happy? Are people in control of their own lives? ..." Humans have complex emotions, have many values—the thousand shards of desire, the godshatter of natural selection. I would say, by the way, that the huge blob of a computation is not just my present terminal values (which I don't really have—I am not a consistent expected utility maximizers); the huge blob of a computation includes the specification of those moral arguments, those justifications, that would sway me if I heard them. So that I can regard my present values, as an approximation to the ideal morality that I would have if I heard all the arguments, to whatever extent such an extrapolation is coherent. [link in the original]

ETA: Just in case you're right and Eliezer somehow meant for that paragraph not to be part of his metaethics, and that his actual metaethics is just "morality is a fixed abstract computation", then I'd ask, "If morality is a fixed abstract computation, then it seems that rationality must also be a fixed abstract computation. But don't you think a complete "solved" metaethics should explain how morality differs from rationality?"

Comment author: gRR 16 April 2012 11:53:58AM 1 point [-]

"If morality is a fixed abstract computation, then it seems that rationality must also be a fixed abstract computation. But don't you think a complete "solved" metaethics should explain how morality differs from rationality?"

Rationality computation outputs statements about the world, morality evaluates them. Rationality is universal and objective, so it is unique as an abstract computation, not just fixed. Morality is arbitrary.

Comment author: Eugine_Nier 17 April 2012 05:06:29AM 3 points [-]

Rationality is universal and objective, so it is unique as an abstract computation, not just fixed. Morality is arbitrary.

How so? Every argument I've heard for why morality is arbitrary applies just as well to rationality.

Comment author: gRR 17 April 2012 11:00:32AM 2 points [-]

If we assume some kind of mathematical realism (which seems to be necessary for "abstract computation" and "uniqueness" to have any meaning) then there exist objectively true statements and computations that generate them. At some point there are Goedelian problems, but at least all of the computations agree on the primitive-recursive truths, which are therefore universal, objective, unique, and true.

Any rational agent (optimization process) in any world with some regularities would exploit these regularities, which means use math. A reflective self-optimizing rational agent would arrive to the same math as us, because the math is unique.

Of course, all these points are made by a fallible human brain and so may be wrong.

But there is nothing even like that for morality. In fact, when a moral statement seems universal under sufficient reflection, it stops being a moral statement and becomes simply rational, like cooperating in the Prisoner's Dilemma when playing against the right opponents.

Comment author: David_Gerard 16 April 2012 10:52:16PM -1 points [-]

Yeah: CEV appears to just move the hard bit. Adding another layer of indirection.

Comment author: Eugine_Nier 15 April 2012 08:15:47PM 2 points [-]

To take Eliezer's statement one meta-level down:

what are the odds that someone who still thought ethics was a Deep Mystery could come up with a correct metaethics?

Comment author: XiXiDu 15 April 2012 11:01:56AM 1 point [-]

Take metaethics, a solved problem: what are the odds that someone who still thought metaethics was a Deep Mystery could write an AI algorithm that could come up with a correct metaethics? I tried that, you know, and in retrospect it didn't work.

What did he mean by "I tried that..."?

Comment author: Will_Newsome 15 April 2012 11:05:29AM 1 point [-]

I'm not at all sure, but I think he means CFAI.

Comment author: Mitchell_Porter 15 April 2012 11:24:48AM 2 points [-]

Possibly he means this.

Comment author: whowhowho 30 January 2013 06:57:46PM 0 points [-]

He may have soleved it, but if only he or someone else could say what the solution was.