Comment author: ChristianKl 07 October 2015 08:52:29AM 0 points [-]

Let's go with this analogy. The good thing to do is ask a variety of experts for safety evaluations, run the code through a wide variety of tests, etc. The think NOT to do is keep the code a secret while looking for mistakes all by yourself.

No, if you don't want to use code you don't give the code to a variety of experts for safety evaluations but you simply don't run the code. Having a public discussion is like running the code untested on a mission critical system.

What utility do you think is gained by discussing the basilisk?

and it's certainly not a good habit.

Strawman. This forum is not a place where things get habitually banned.

Comment author: anon85 07 October 2015 05:40:23PM 0 points [-]

What utility do you think is gained by discussing the basilisk?

An interesting discussion that leads to better understanding of decision theories? Like, the same utility as is gained by any other discussion on LW, pretty much.

Strawman. This forum is not a place where things get habitually banned.

Sure, but you're the one that was going on about the importance of the mindset and culture; since you brought it up in the context of banning discussion, it sounded like you were saying that such censorship was part of a mindset/culture that you approve of.

Comment author: RichardKennaway 07 October 2015 08:18:50AM 3 points [-]

Somehow, blackmail from the future seems less plausible to me than every single one of your examples. Not sure why exactly.

How plausible do you find TDT and related decision theories as normative accounts of decision making, or at least as work towards such accounts? They open whole new realms of situations like Pascal's Mugging, of which Roko's Basilisk is one. If you're going to think in detail about such decision theories, and adopt one as normative, you need to have an answer to these situations.

Once you've decided to study something seriously, the plausibility heuristic is no longer available.

Comment author: anon85 07 October 2015 05:35:06PM 1 point [-]

I find TDT to be basically bullshit except possibly when it is applied to entities which literally see each others' code, in which case I'm not sure (I'm not even sure if the concept of "decision" even makes sense in that case).

I'd go so far as to say that anyone who advocates cooperating in a one-shot prisoners' dilemma simply doesn't understand the setting. By definition, defecting gives you a better outcome than cooperating. Anyone who claims otherwise is changing the definition of the prisoners' dilemma.

Comment author: RichardKennaway 07 October 2015 11:58:47AM 2 points [-]

If a philosophical framework causes you to accept a basilisk, I view that as grounds for rejecting the framework, not for accepting the basilisk.

...

To state it yet another way: to me, the basilisk has the same status as an ontological argument for God. Even if I can't find the flaw in the argument, I'm confident in rejecting it anyway.

Despite the other things I've said here, that is my attitude as well. But I recognise that when I take that attitude, I am not solving the problem, only ignoring it. It may be perfectly sensible to ignore a problem, even a serious one (comparative advantage etc.). But dissolving a paradox is not achieved by clinging to one of the conflicting thoughts and ignoring the others. (Bullet-swallowing seems to consist of seizing onto the most novel one.) Eliminating the paradox requires showing where and how the thoughts went wrong.

Comment author: anon85 07 October 2015 05:29:14PM 1 point [-]

I agree that resolving paradoxes is an important intellectual exercise, and that I wouldn't be satisfied with simply ignoring an ontological argument (I'd want to find the flaw). But the best way to find such flaws is to discuss the ideas with others. At no point should one assign such a high probability to ideas like Roko's basilisk being actually sound that one refuses to discuss them with others.

Comment author: hairyfigment 07 October 2015 07:22:15AM 1 point [-]

I believe he thinks that sufficiently clever idiots competing to shoot off their own feet will find some way to do so.

Comment author: anon85 07 October 2015 05:24:57PM 0 points [-]

It seems unlikely that they would, if their gun is some philosophical decision theory stuff about blackmail from their future. I don't expect that gun to ever fire, no matter how many times you click the trigger.

Comment author: RobbBB 07 October 2015 07:17:53AM *  2 points [-]

Sorry, I'll be more concrete; "there's a serious risk" is really vague wording. What would surprise me greatly is if I heard that Eliezer assigned even a 5% probability to there being a realistic quick fix to Roko's argument that makes it work on humans. I think a larger reason for the ban was just that Eliezer was angry with Roko for trying to spread what Roko thought was an information hazard, and angry people lash out (even when it doesn't make a ton of strategic sense).

Comment author: anon85 07 October 2015 05:23:17PM 1 point [-]

Probably not a quick fix, but I would definitely say Eliezer gives significant chances (say, 10%) to there being some viable version of the Basilisk, which is why he actively avoids thinking about it.

If Eliezer was just angry at Roko, he would have yelled or banned Roko; instead, he banned all discussion of the subject. That doesn't even make sense as a "slashing out" reaction against Roko.

Comment author: RichardKennaway 06 October 2015 06:47:35PM 3 points [-]

The basilisk is so a priori implausible that you should be extremely suspicious of fancy arguments claiming to prove it.

So are: God, superintelligent AI, universal priors, radical life extension, and any really big idea whatever; as well as the impossibility of each of these.

Plausibility is fine as a screening process for deciding where you're going to devote your efforts, but terrible as an epistemological tool.

Comment author: anon85 07 October 2015 12:57:33AM *  0 points [-]

Somehow, blackmail from the future seems less plausible to me than every single one of your examples. Not sure why exactly.

Comment author: ChristianKl 06 October 2015 06:54:42PM 2 points [-]

Eliezer thought it was so plausible that he banned discussion of it

If you are a programmer and think your code is safe because you see no way things could go wrong, it's still not good to believe that it isn't plausible that there's a security hole in your code.

You rather practice defense in depth and plan for the possibility that things can go wrong somewhere in your code, so you add safety precautions. Even when there isn't what courts call reasonable doubt a good safety engineer still adds additional safety procautions in security critical code. Eliezer deals with FAI safety. As a result it's good for him to have mindset of really caring about safety.

German nuclear power station have trainings for their desk workers to teach the desk workers to not cut themselves with paper. That alone seems strange to outsiders but everyone in Germany thinks that it's very important for nuclear power stations to foster a culture of safety even when that means something going overboard.

Comment author: anon85 07 October 2015 12:56:38AM 1 point [-]

If you are a programmer and think your code is safe because you see no way things could go wrong, it's still not good to believe that it isn't plausible that there's a security hole in your code.

Let's go with this analogy. The good thing to do is ask a variety of experts for safety evaluations, run the code through a wide variety of tests, etc. The think NOT to do is keep the code a secret while looking for mistakes all by yourself. If you keep your code out of the public domain, it is more likely to have security issues, since it was not scrutinized by the public. Banning discussion is almost never correct, and it's certainly not a good habit.

Comment author: ChristianKl 06 October 2015 06:54:03PM 0 points [-]

The basilisk is so a priori implausible that you should be extremely suspicious of fancy arguments claiming to prove it.

Finding an idea plausible has little to do with being extremely suspicious of fancy arguments claiming to prove it.

Idea that aren't proven to be impossible are plausible even when there are no convincing arguments in favor of them.

Comment author: anon85 07 October 2015 12:50:10AM 1 point [-]

Ideas that aren't proven to be impossible are possible. They don't have to be plausible.

Comment author: RobbBB 06 October 2015 11:29:20PM *  0 points [-]

If you're saying 'LessWrongers think there's a serious risk they'll be acausally blackmailed by a rogue AI', then that seems to be false. That even seems to be false in Eliezer's case, and Eliezer definitely isn't 'LessWrong'. If you're saying 'LessWrongers think acausal trade in general is possible,' then that seems true but I don't see why that's ridiculous.

Is there something about acausal trade in general that you're objecting to, beyond the specific problems with Roko's argument?

Comment author: anon85 07 October 2015 12:44:15AM 0 points [-]

That even seems to be false in Eliezer's case, and Eliezer definitely isn't 'LessWrong'.

It seems we disagree on this factual issue. Eliezer does think there is a risk of acausal blackmail, or else he wouldn't have banned discussion of it.

Comment author: Viliam 06 October 2015 03:06:03PM *  2 points [-]

This is evidence that Yudkowsky believed (...) that at least it was plausible enough that could be developed in a correct argument, and he was genuinely scared by it.

Just to be sure, since you seem to disagree with this opinion (whether it is actually Yudkowsky's opinion or not), what exactly is it that you believe?

a) There is absolutely no way one could be harmed by thinking about not-yet-existing dangerous entities; even if those entities in the future will be able to learn about the fact that the person was thinking about them in this specific way.

b) There is a way one could be harmed by thinking about not-yet-existing dangerous entities, but the way to do this is completely different from what Roko proposed.

If it happens to be (b), then it still makes sense to be angry about publicly opening the whole topic of "let's use our intelligence to discover the thoughts that may harm us by us thinking about them -- and let's do it in a public forum where people are interested in decision theories, so they are more qualified than average to find the right answer." Even if the proper way to harm oneself is different from what Roko proposed, making this a publicly debated topic increases the chance of someone finding the correct solution. The problem is not the proposed basilisk, but rather inviting people to compete in clever self-harm; especially the kind of people known for being hardly able to resist such invitation.

Comment author: anon85 06 October 2015 05:38:34PM *  0 points [-]

I'm not the person you replied to, but I mostly agree with (a) and reject (b). There's no way you can could possibly know enough about a not-yet-existing entity to understand any of its motivations; the entities that you're thinking about and the entities that will exist in the future are not even close to the same. I outlined some more thoughts here.

View more: Prev | Next