XiXiDu comments on Open Thread: July 2010, Part 2 - Less Wrong

6 Post author: Alicorn 09 July 2010 06:54AM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (770)

You are viewing a single comment's thread. Show more comments above.

Comment author: XiXiDu 06 August 2010 10:40:08AM *  13 points [-]

I know, it wasn't my intention to discredit Peer, I quite like his ideas. I'm probably more crazy than him anyway.

But if I can come up with such conclusions, who else will? Also, why isn't anyone out to kill people, or will be? I'm serious, why not? Just imagine EY found out that we can be reasonable sure, for example Google, would let loose a rogue AI soon. Given how the LW audience is inclined to act upon 'mere' probability estimates, how wouldn't it be appropriate to bomb Google, given that was the only way to stop them in due time from turning the world into a living hell? And how isn't this meme, given the right people and circumstances, a great danger? Sure, me saying EY might be a greater danger was nonsense, just said to provoke some response. By definition, not much could be worse than uFAI.

This incident is simply a good situation to extrapolate. If a thought-experiment can be deemed to be dangerous enough to be not just censored and deleted but for people to be told not even to seek any knowledge of it, much less discuss it, I'm wondering about the possible reaction to a imminent and tangible danger.

Comment deleted 23 September 2010 05:21:40AM [-]
Comment deleted 24 September 2010 06:42:21AM [-]
Comment deleted 24 September 2010 06:59:47AM [-]
Comment deleted 06 August 2010 11:40:19AM *  [-]
Comment deleted 06 August 2010 01:49:02PM [-]
Comment deleted 06 August 2010 01:56:18PM [-]
Comment deleted 06 August 2010 10:02:39PM [-]
Comment deleted 06 August 2010 11:58:26AM *  [-]
Comment author: Blueberry 06 August 2010 12:32:53PM *  7 points [-]

Heh, that makes Roko's scenario similar to the Missionary Paradox: if only those who know about God but don't believe go to hell, it's harmful to spread the idea of God. (As I understand it, this doesn't come up because most missionaries think you'll go to hell even if you don't know about the idea of God.)

But I don't think any God is supposed to follow a human CEV; most religions seem to think it's the other way around.

Comment author: NancyLebovitz 06 August 2010 02:00:27PM 4 points [-]

The more recent analysis I've read says that people pretty much become suicide bombers for nationalist reasons, not religious reasons.

I suppose that "There should not be American bases on the sacred soil of Saudi Arabia" is a hybrid of the two, and so might be "I wanted to kill because Muslims were being hurt"-- it's a matter of group identity more than "Allah wants it".

I don't have specifics for the 9/11 bombers.

Comment author: katydee 06 August 2010 12:10:55PM 0 points [-]

Either I misunderstand CEV, or the above statement re: the Abrahamic god following CEV is false.

Comment author: XiXiDu 06 August 2010 12:19:37PM 2 points [-]

Coherent Extrapolated Volition

Long distance: An extrapolated volition your present-day self finds incomprehensible; not outrageous or annoying, but blankly incomprehensible.

This is exactly the argument religious people use to excuse any shortcomings of their personal FAI. Namely, their personal FAI knows better than you what's best for your AND everyone else.

What average people do is follow what is being taught here on LW. They decide based on their prior. Their probability estimates tell them that their FAI is likely to exist and make up excuses for extraordinary decisions based on the possible existence of it. That is, support their FAI while trying to inhibit other uFAI all in the best interest of the world at large.

Comment author: orthonormal 06 August 2010 09:57:49PM 1 point [-]

Yahweh and the associated moral system are far from incomprehensible if you know the cultural context of the Israelites. It's a recognizably human morality, just a brutal one obsessed with purity of various sorts.

Comment author: XiXiDu 07 August 2010 09:17:32AM *  6 points [-]

It is not about the moral system being incomprehensible but the acts of the FAI. Whenever something bad happens religious people excuse it with an argument based on "higher intention". This is the gist of what I wanted to highlight. The similarity between religious people and those true believers into the technological singularity and AI's. This is not to say it is the same. I'm not arguing about that. I'm saying that this might draw the same kind of people committing the same kind of atrocities. This is very dangerous.

If people don't like anything happening, i.e. don't understand it, it's claimed to be a means to an end that will ultimately benefit their extrapolated volition.

People are not going to claim this in public. But I know that there are people here on LW who are disposed to extensive violence if necessary.

To be clear, I do not doubt the possibilities talked about on LW. I'm not saying they are nonsense like the old religions. What I'm on about is that the ideas the SIAI is based on, while not being nonsense, are posed to draw the same fanatic fellowship and cause the same extreme decisions.

Ask yourself, wouldn't you fly a plane into a tower if that was the only way to disable Skynet? The difference between religion and the risk of uFAI makes it even more dangerous. This crowd is actually highly intelligent and their incentive based on more than fairy tales told by goatherders. And if dumb people are already able to commit large-scale atrocities based on such nonsense, what are a bunch of highly-intelligent and devoted geeks who see a tangible danger able and willing to do? More so as in this case the very same people who believe it are the ones who think they must act themselves because their God doesn't even exist yet.

Comment author: orthonormal 07 August 2010 02:34:40PM 8 points [-]

Ask yourself, wouldn't you fly a plane into a tower if that was the only way to disable Skynet?

Yes. I would also drop a nuke on New York if it were the only way to prevent global nuclear war. These are both extremely unlikely scenarios.

It's very correct to be suspicious of claims that the stakes are that high, given that irrational memes have a habit of postulating such high stakes. However, assuming thereby that the stakes never could actually be that high, regardless of the evidence, is another way of shooting yourself in the foot.

Comment author: katydee 06 August 2010 09:50:37PM 1 point [-]

The link and quotation you posted do not seem to back up your argument that the Abrahamic god follows CEV. Could you clarify?

Comment author: XiXiDu 07 August 2010 08:49:02AM 2 points [-]

It's not about it following CEV but people believing it, that it acts in their best interest. Reasons are subordinate. It is the similar systematic of positive and negative incentive that I wanted to highlight.

I grew up in a family of Jehovah's Witnesses. I can assure you that all believed this to be the case.

Faith is considered the way to happiness.

Positive incentive:

“O you who believe! What is the matter with you, that, when you are asked to go forth in the Cause of Allâh, you cling heavily to the earth? Do you prefer the life of this world to the Hereafter? But little is the comfort of this life, as compared with the Hereafter.”

Whoever does right, whether male or female, and is a believer, We will make him live a good life, and We will award them their reward for the best of what they used to do. (Quran, 16:97)

Negative incentive:

"You dissipated the good things you had in your worldly life and enjoyed yourself in it. So today you are being repaid with the punishment of humiliation for being arrogant in the Earth without any right and for being deviators." (Surat al-Ahqaf: 20)

I could find heaps of arguments for Christianity that highlight the same believe of God knowing what's the best for you and the world. This is what most people on this planet believe and this is also the underpinning of the rapture of the nerds.

Comment author: katydee 07 August 2010 10:06:15AM 1 point [-]

Ah, I understand-- except that I think the "negative incentive" element we're discussing is absurd, would obviously trigger failsafes with CEV as described, etc.

Comment author: XiXiDu 07 August 2010 10:46:27AM *  2 points [-]

There'll always be elements that suffer, that is perceive FAI as uFAI subjectively.

Comment author: Blueberry 06 August 2010 12:30:14PM 0 points [-]

If an AI does what Roko suggested, it's not friendly. We don't know what, if anything, CEV will output, but I don't see any reason to think CEV would enact Roko's scenario.

Comment author: cousin_it 06 August 2010 12:33:45PM *  0 points [-]

Roko thinks (or thought) it would. I do too. Can't argue it in detail here, sorry.

Comment author: [deleted] 03 March 2012 10:26:01PM 0 points [-]

Until about a month ago, I would have agreed, but some posts I have since read on LW made me update the probability of CEV wanting that upwards.

Comment author: Blueberry 25 March 2012 01:20:26AM 0 points [-]

Really, please explain (or PM me if it would require breaking the gag rule on Roko's scenario). Why would CEV want that?

Comment author: Blueberry 25 March 2012 12:54:50AM -2 points [-]

Really, please explain (or PM me if it would require breaking the gag rule on Roko's scenario). Why would CEV want that?

Comment author: wedrifid 25 March 2012 02:40:58AM *  1 point [-]

Why would CEV want that?

Because 'CEV' must be instantiated on a group of agents (usually humans). Some humans are assholes. So for some value of aGroup, CEV<aGroup> does assholish things. Hopefully the group of all humans doesn't create a CEV that makes FAI<CEV<all humans>> an outright uFAI from our perspective but we certainly shouldn't count on it.

Comment author: Blueberry 25 March 2012 03:18:10AM *  0 points [-]

Some humans are assholes. So for some value of aGroup, CEV<aGroup> does assholish things.

That's not necessarily true. CEV isn't precisely defined but it's intended to represent the idealized version of our desires and meta-desires. So even if we take a group of assholes, they don't necessarily want to be assholes, or want to want to be assholes, or maybe they wouldn't want to if they knew more and were smarter.

Comment author: wedrifid 25 March 2012 06:59:22AM *  2 points [-]

I refer, of course, to people whose preferences really are different to our own. Coherent Extrapolated Assholes. I don't refer to people who would really have preferences that I would consider acceptable if they just knew a bit more.

You asked for an explanation of how a correctly implemented 'CEV' could want something abhorrent. That's how.

There is an unfortunate tendency to glorify the extrapolation process and pretend that it makes any given individual or group have acceptable values. It need not.

Comment author: [deleted] 25 March 2012 10:21:01AM *  1 point [-]

Upvoted for the phrase “Coherent Extrapolated Assholes”. Best. Insult. Ever.

Seriously, though, I don't think there are many CEAs around, anyway. (This doesn't mean there are none, either. (I was going to link to this as an example of one, but I'm not sure Hitler would have done what he did had he known about late-20th-century results about heterosis, Ashkenazi Jew intelligence, etc.)) This mean that I think it's very, very unlikely for CEV<aGroup> to be evil<army1987> (and even less likely to be evil<CEV<army1987>>), unless the membership criteria to aGroup are gerrymandered to make it so.

Comment author: Vaniver 25 March 2012 07:22:45AM 0 points [-]

There is an unfortunate tendency to glorify the extrapolation process and pretend that it makes any given individual or group have acceptable values. It need not.

It seemed odd to me that so few people were bothered by the claims that CEV shouldn't care much about the inputs. If you expect it to give similar results if you put in a chimpanzee and a murderer and Archimedes, then why put in anything at all instead of just printing out the only results it gives?

Comment deleted 25 March 2012 01:25:59AM *  [-]
Comment deleted 25 March 2012 01:31:07AM [-]
Comment author: TimS 25 March 2012 02:58:17AM 0 points [-]

If you believe in moral progress (and CEV seems to rely on that position), then there's every reason to think that future-society would want to make changes to how we live, if future-society had the capacity to make that type of intervention.

In short, wouldn't you change the past to prevent the occurrence of chattel slavery if you could? (If you don't like that example, substitute preventing the October revolution or whatever example fits your preferences).

Comment author: Blueberry 25 March 2012 03:15:26AM 0 points [-]

I wouldn't torture innocent people to prevent it, no.

Comment author: TimS 25 March 2012 03:42:46AM 0 points [-]

Punishment from the future is spooky enough. Imagine what an anti-Guns of the South would be like for the temporal locals. Not pleasant, that's for sure.

Comment author: wedrifid 25 March 2012 07:34:30AM -1 points [-]

If you believe in moral progress (and CEV seems to rely on that position)

It's more agnostic on the issue. It works just as well for the ultimate conservative.

Comment author: TheOtherDave 25 March 2012 04:24:45PM 0 points [-]

Doesn't CEV implicitly assert that there exists a set of moral assertions M that is more reliably moral than anything humans assert today, and that it's possible for a sufficiently intelligent system to derive M?

That sure sounds like a belief in moral progress to me.

Granted, it doesn't imply that humans left to their own devices will achieve moral progress. But the same is true of technological progress.

Comment author: wedrifid 25 March 2012 04:28:19PM 1 point [-]

Doesn't CEV implicitly assert that there exists a set of moral assertions M that is more reliably moral than anything humans assert today, and that it's possible for a sufficiently intelligent system to derive M?

The implicit assertion is "Greater or Equal", not "Greater".

Run on a True Conservative it will return the morals that the conservative currently has.

Comment author: TheOtherDave 25 March 2012 07:01:52PM 0 points [-]

Mm.
I'll certainly agree that anyone for whom that's true deserves the title "True Conservative."

I don't think I've ever met anyone who meets that description, though I've met people who would probably describe themselves that way.

Presumably, someone who believes this is true of themselves would consider the whole notion of extrapolating the target definition for a superhumanly powerful optimization process to be silly, though, and consider the label CEV to be technically accurate, in the same sense that I'm currently extrapolating the presence of my laptop, but to imply falsehoods.