Blueberry comments on Open Thread: July 2010, Part 2 - Less Wrong

6 Post author: Alicorn 09 July 2010 06:54AM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (770)

You are viewing a single comment's thread. Show more comments above.

Comment author: Blueberry 06 August 2010 12:30:14PM 0 points [-]

If an AI does what Roko suggested, it's not friendly. We don't know what, if anything, CEV will output, but I don't see any reason to think CEV would enact Roko's scenario.

Comment author: cousin_it 06 August 2010 12:33:45PM *  0 points [-]

Roko thinks (or thought) it would. I do too. Can't argue it in detail here, sorry.

Comment author: [deleted] 03 March 2012 10:26:01PM 0 points [-]

Until about a month ago, I would have agreed, but some posts I have since read on LW made me update the probability of CEV wanting that upwards.

Comment author: Blueberry 25 March 2012 01:20:26AM 0 points [-]

Really, please explain (or PM me if it would require breaking the gag rule on Roko's scenario). Why would CEV want that?

Comment author: Blueberry 25 March 2012 12:54:50AM -2 points [-]

Really, please explain (or PM me if it would require breaking the gag rule on Roko's scenario). Why would CEV want that?

Comment author: wedrifid 25 March 2012 02:40:58AM *  1 point [-]

Why would CEV want that?

Because 'CEV' must be instantiated on a group of agents (usually humans). Some humans are assholes. So for some value of aGroup, CEV<aGroup> does assholish things. Hopefully the group of all humans doesn't create a CEV that makes FAI<CEV<all humans>> an outright uFAI from our perspective but we certainly shouldn't count on it.

Comment author: Blueberry 25 March 2012 03:18:10AM *  0 points [-]

Some humans are assholes. So for some value of aGroup, CEV<aGroup> does assholish things.

That's not necessarily true. CEV isn't precisely defined but it's intended to represent the idealized version of our desires and meta-desires. So even if we take a group of assholes, they don't necessarily want to be assholes, or want to want to be assholes, or maybe they wouldn't want to if they knew more and were smarter.

Comment author: wedrifid 25 March 2012 06:59:22AM *  2 points [-]

I refer, of course, to people whose preferences really are different to our own. Coherent Extrapolated Assholes. I don't refer to people who would really have preferences that I would consider acceptable if they just knew a bit more.

You asked for an explanation of how a correctly implemented 'CEV' could want something abhorrent. That's how.

There is an unfortunate tendency to glorify the extrapolation process and pretend that it makes any given individual or group have acceptable values. It need not.

Comment author: [deleted] 25 March 2012 10:21:01AM *  1 point [-]

Upvoted for the phrase “Coherent Extrapolated Assholes”. Best. Insult. Ever.

Seriously, though, I don't think there are many CEAs around, anyway. (This doesn't mean there are none, either. (I was going to link to this as an example of one, but I'm not sure Hitler would have done what he did had he known about late-20th-century results about heterosis, Ashkenazi Jew intelligence, etc.)) This mean that I think it's very, very unlikely for CEV<aGroup> to be evil<army1987> (and even less likely to be evil<CEV<army1987>>), unless the membership criteria to aGroup are gerrymandered to make it so.

Comment author: Vaniver 25 March 2012 07:22:45AM 0 points [-]

There is an unfortunate tendency to glorify the extrapolation process and pretend that it makes any given individual or group have acceptable values. It need not.

It seemed odd to me that so few people were bothered by the claims that CEV shouldn't care much about the inputs. If you expect it to give similar results if you put in a chimpanzee and a murderer and Archimedes, then why put in anything at all instead of just printing out the only results it gives?

Comment deleted 25 March 2012 01:25:59AM *  [-]
Comment deleted 25 March 2012 01:31:07AM [-]
Comment author: TimS 25 March 2012 02:58:17AM 0 points [-]

If you believe in moral progress (and CEV seems to rely on that position), then there's every reason to think that future-society would want to make changes to how we live, if future-society had the capacity to make that type of intervention.

In short, wouldn't you change the past to prevent the occurrence of chattel slavery if you could? (If you don't like that example, substitute preventing the October revolution or whatever example fits your preferences).

Comment author: Blueberry 25 March 2012 03:15:26AM 0 points [-]

I wouldn't torture innocent people to prevent it, no.

Comment author: TimS 25 March 2012 03:42:46AM 0 points [-]

Punishment from the future is spooky enough. Imagine what an anti-Guns of the South would be like for the temporal locals. Not pleasant, that's for sure.

Comment author: wedrifid 25 March 2012 07:34:30AM -1 points [-]

If you believe in moral progress (and CEV seems to rely on that position)

It's more agnostic on the issue. It works just as well for the ultimate conservative.

Comment author: TheOtherDave 25 March 2012 04:24:45PM 0 points [-]

Doesn't CEV implicitly assert that there exists a set of moral assertions M that is more reliably moral than anything humans assert today, and that it's possible for a sufficiently intelligent system to derive M?

That sure sounds like a belief in moral progress to me.

Granted, it doesn't imply that humans left to their own devices will achieve moral progress. But the same is true of technological progress.

Comment author: wedrifid 25 March 2012 04:28:19PM 1 point [-]

Doesn't CEV implicitly assert that there exists a set of moral assertions M that is more reliably moral than anything humans assert today, and that it's possible for a sufficiently intelligent system to derive M?

The implicit assertion is "Greater or Equal", not "Greater".

Run on a True Conservative it will return the morals that the conservative currently has.

Comment author: TheOtherDave 25 March 2012 07:01:52PM 0 points [-]

Mm.
I'll certainly agree that anyone for whom that's true deserves the title "True Conservative."

I don't think I've ever met anyone who meets that description, though I've met people who would probably describe themselves that way.

Presumably, someone who believes this is true of themselves would consider the whole notion of extrapolating the target definition for a superhumanly powerful optimization process to be silly, though, and consider the label CEV to be technically accurate, in the same sense that I'm currently extrapolating the presence of my laptop, but to imply falsehoods.