XiXiDu comments on Open Thread: July 2010, Part 2 - Less Wrong
You are viewing a comment permalink. View the original post to see all comments and the full post content.
You are viewing a comment permalink. View the original post to see all comments and the full post content.
Comments (770)
I know, it wasn't my intention to discredit Peer, I quite like his ideas. I'm probably more crazy than him anyway.
But if I can come up with such conclusions, who else will? Also, why isn't anyone out to kill people, or will be? I'm serious, why not? Just imagine EY found out that we can be reasonable sure, for example Google, would let loose a rogue AI soon. Given how the LW audience is inclined to act upon 'mere' probability estimates, how wouldn't it be appropriate to bomb Google, given that was the only way to stop them in due time from turning the world into a living hell? And how isn't this meme, given the right people and circumstances, a great danger? Sure, me saying EY might be a greater danger was nonsense, just said to provoke some response. By definition, not much could be worse than uFAI.
This incident is simply a good situation to extrapolate. If a thought-experiment can be deemed to be dangerous enough to be not just censored and deleted but for people to be told not even to seek any knowledge of it, much less discuss it, I'm wondering about the possible reaction to a imminent and tangible danger.
Heh, that makes Roko's scenario similar to the Missionary Paradox: if only those who know about God but don't believe go to hell, it's harmful to spread the idea of God. (As I understand it, this doesn't come up because most missionaries think you'll go to hell even if you don't know about the idea of God.)
But I don't think any God is supposed to follow a human CEV; most religions seem to think it's the other way around.
The more recent analysis I've read says that people pretty much become suicide bombers for nationalist reasons, not religious reasons.
I suppose that "There should not be American bases on the sacred soil of Saudi Arabia" is a hybrid of the two, and so might be "I wanted to kill because Muslims were being hurt"-- it's a matter of group identity more than "Allah wants it".
I don't have specifics for the 9/11 bombers.
Either I misunderstand CEV, or the above statement re: the Abrahamic god following CEV is false.
Coherent Extrapolated Volition
This is exactly the argument religious people use to excuse any shortcomings of their personal FAI. Namely, their personal FAI knows better than you what's best for your AND everyone else.
What average people do is follow what is being taught here on LW. They decide based on their prior. Their probability estimates tell them that their FAI is likely to exist and make up excuses for extraordinary decisions based on the possible existence of it. That is, support their FAI while trying to inhibit other uFAI all in the best interest of the world at large.
Yahweh and the associated moral system are far from incomprehensible if you know the cultural context of the Israelites. It's a recognizably human morality, just a brutal one obsessed with purity of various sorts.
It is not about the moral system being incomprehensible but the acts of the FAI. Whenever something bad happens religious people excuse it with an argument based on "higher intention". This is the gist of what I wanted to highlight. The similarity between religious people and those true believers into the technological singularity and AI's. This is not to say it is the same. I'm not arguing about that. I'm saying that this might draw the same kind of people committing the same kind of atrocities. This is very dangerous.
If people don't like anything happening, i.e. don't understand it, it's claimed to be a means to an end that will ultimately benefit their extrapolated volition.
People are not going to claim this in public. But I know that there are people here on LW who are disposed to extensive violence if necessary.
To be clear, I do not doubt the possibilities talked about on LW. I'm not saying they are nonsense like the old religions. What I'm on about is that the ideas the SIAI is based on, while not being nonsense, are posed to draw the same fanatic fellowship and cause the same extreme decisions.
Ask yourself, wouldn't you fly a plane into a tower if that was the only way to disable Skynet? The difference between religion and the risk of uFAI makes it even more dangerous. This crowd is actually highly intelligent and their incentive based on more than fairy tales told by goatherders. And if dumb people are already able to commit large-scale atrocities based on such nonsense, what are a bunch of highly-intelligent and devoted geeks who see a tangible danger able and willing to do? More so as in this case the very same people who believe it are the ones who think they must act themselves because their God doesn't even exist yet.
Yes. I would also drop a nuke on New York if it were the only way to prevent global nuclear war. These are both extremely unlikely scenarios.
It's very correct to be suspicious of claims that the stakes are that high, given that irrational memes have a habit of postulating such high stakes. However, assuming thereby that the stakes never could actually be that high, regardless of the evidence, is another way of shooting yourself in the foot.
The link and quotation you posted do not seem to back up your argument that the Abrahamic god follows CEV. Could you clarify?
It's not about it following CEV but people believing it, that it acts in their best interest. Reasons are subordinate. It is the similar systematic of positive and negative incentive that I wanted to highlight.
I grew up in a family of Jehovah's Witnesses. I can assure you that all believed this to be the case.
Faith is considered the way to happiness.
Positive incentive:
Negative incentive:
I could find heaps of arguments for Christianity that highlight the same believe of God knowing what's the best for you and the world. This is what most people on this planet believe and this is also the underpinning of the rapture of the nerds.
Ah, I understand-- except that I think the "negative incentive" element we're discussing is absurd, would obviously trigger failsafes with CEV as described, etc.
There'll always be elements that suffer, that is perceive FAI as uFAI subjectively.
If an AI does what Roko suggested, it's not friendly. We don't know what, if anything, CEV will output, but I don't see any reason to think CEV would enact Roko's scenario.
Roko thinks (or thought) it would. I do too. Can't argue it in detail here, sorry.
Until about a month ago, I would have agreed, but some posts I have since read on LW made me update the probability of CEV wanting that upwards.
Really, please explain (or PM me if it would require breaking the gag rule on Roko's scenario). Why would CEV want that?
Really, please explain (or PM me if it would require breaking the gag rule on Roko's scenario). Why would CEV want that?
Because 'CEV' must be instantiated on a group of agents (usually humans). Some humans are assholes. So for some value of aGroup, CEV<aGroup> does assholish things. Hopefully the group of all humans doesn't create a CEV that makes FAI<CEV<all humans>> an outright uFAI from our perspective but we certainly shouldn't count on it.
That's not necessarily true. CEV isn't precisely defined but it's intended to represent the idealized version of our desires and meta-desires. So even if we take a group of assholes, they don't necessarily want to be assholes, or want to want to be assholes, or maybe they wouldn't want to if they knew more and were smarter.
I refer, of course, to people whose preferences really are different to our own. Coherent Extrapolated Assholes. I don't refer to people who would really have preferences that I would consider acceptable if they just knew a bit more.
You asked for an explanation of how a correctly implemented 'CEV' could want something abhorrent. That's how.
There is an unfortunate tendency to glorify the extrapolation process and pretend that it makes any given individual or group have acceptable values. It need not.
Upvoted for the phrase “Coherent Extrapolated Assholes”. Best. Insult. Ever.
Seriously, though, I don't think there are many CEAs around, anyway. (This doesn't mean there are none, either. (I was going to link to this as an example of one, but I'm not sure Hitler would have done what he did had he known about late-20th-century results about heterosis, Ashkenazi Jew intelligence, etc.)) This mean that I think it's very, very unlikely for CEV<aGroup> to be evil<army1987> (and even less likely to be evil<CEV<army1987>>), unless the membership criteria to aGroup are gerrymandered to make it so.
It seemed odd to me that so few people were bothered by the claims that CEV shouldn't care much about the inputs. If you expect it to give similar results if you put in a chimpanzee and a murderer and Archimedes, then why put in anything at all instead of just printing out the only results it gives?
If you believe in moral progress (and CEV seems to rely on that position), then there's every reason to think that future-society would want to make changes to how we live, if future-society had the capacity to make that type of intervention.
In short, wouldn't you change the past to prevent the occurrence of chattel slavery if you could? (If you don't like that example, substitute preventing the October revolution or whatever example fits your preferences).
I wouldn't torture innocent people to prevent it, no.
Punishment from the future is spooky enough. Imagine what an anti-Guns of the South would be like for the temporal locals. Not pleasant, that's for sure.
It's more agnostic on the issue. It works just as well for the ultimate conservative.
Doesn't CEV implicitly assert that there exists a set of moral assertions M that is more reliably moral than anything humans assert today, and that it's possible for a sufficiently intelligent system to derive M?
That sure sounds like a belief in moral progress to me.
Granted, it doesn't imply that humans left to their own devices will achieve moral progress. But the same is true of technological progress.
The implicit assertion is "Greater or Equal", not "Greater".
Run on a True Conservative it will return the morals that the conservative currently has.
Mm.
I'll certainly agree that anyone for whom that's true deserves the title "True Conservative."
I don't think I've ever met anyone who meets that description, though I've met people who would probably describe themselves that way.
Presumably, someone who believes this is true of themselves would consider the whole notion of extrapolating the target definition for a superhumanly powerful optimization process to be silly, though, and consider the label CEV to be technically accurate, in the same sense that I'm currently extrapolating the presence of my laptop, but to imply falsehoods.