Comment author: rayalez 06 October 2015 06:45:18PM *  2 points [-]

I'm new to the subject, so I'm sorry if the following is obvious or completely wrong, but the comment left by Eliezer doesn't seem like something that would be written by a smart person who is trying to suppress information. I seriously doubt that EY didn't know about Streisand effect.

However the comment does seem like something that would be written by a smart person who is trying to create a meme or promote his blog.

In HPMOR characters give each other advice "to understand a plot, assume that what happened was the intended result, and look at who benefits." The idea of Roko's basilisk went viral and lesswrong.com got a lot of traffic from popular news sites(I'm assuming).

I also don't think that there's anything wrong with it, I'm just sayin'.

Comment author: RobbBB 06 October 2015 11:13:18PM *  4 points [-]

The line goes "to fathom a strange plot, one technique was to look at what ended up happening, assume it was the intended result, and ask who benefited". But in the real world strange secret complicated Machiavellian plots are pretty rare, and successful strange secret complicated Machiavellian plots are even rarer. So I'd be wary of applying this rule to explain big once-off events outside of fiction. (Even to HPMoR's author!)

I agree Eliezer didn't seem to be trying very hard to suppress information. I think that's probably just because he's a human, and humans get angry when they see other humans defecting from a (perceived) social norm, and anger plus time pressure causes hasty dumb decisions. I don't think this is super complicated. Though I hope he'd have acted differently if he thought the infohazard risk was really severe, as opposed to just not-vanishingly-small.

Comment author: ChristianKl 06 October 2015 06:54:42PM 2 points [-]

Eliezer thought it was so plausible that he banned discussion of it

If you are a programmer and think your code is safe because you see no way things could go wrong, it's still not good to believe that it isn't plausible that there's a security hole in your code.

You rather practice defense in depth and plan for the possibility that things can go wrong somewhere in your code, so you add safety precautions. Even when there isn't what courts call reasonable doubt a good safety engineer still adds additional safety procautions in security critical code. Eliezer deals with FAI safety. As a result it's good for him to have mindset of really caring about safety.

German nuclear power station have trainings for their desk workers to teach the desk workers to not cut themselves with paper. That alone seems strange to outsiders but everyone in Germany thinks that it's very important for nuclear power stations to foster a culture of safety even when that means something going overboard.

Comment author: RobbBB 06 October 2015 08:32:16PM *  0 points [-]
Comment author: philh 06 October 2015 02:16:18PM 8 points [-]

There are lots of good reasons Eliezer shouldn't have banned Roko

IIRC, Eliezer didn't ban Roko, just discussion of the basilisk, and Roko deleted his account shortly afterwards.

Comment author: RobbBB 06 October 2015 07:00:32PM 1 point [-]

Thanks, fixed!

Comment author: V_V 06 October 2015 10:51:51AM *  2 points [-]

When Roko posted about the Basilisk, I very foolishly yelled at him, called him an idiot, and then deleted the post. [...] Why I yelled at Roko: Because I was caught flatfooted in surprise, because I was indignant to the point of genuine emotional shock, at the concept that somebody who thought they'd invented a brilliant idea that would cause future AIs to torture people who had the thought, had promptly posted it to the public Internet. In the course of yelling at Roko to explain why this was a bad thing, I made the further error---keeping in mind that I had absolutely no idea that any of this would ever blow up the way it did, if I had I would obviously have kept my fingers quiescent---of not making it absolutely clear using lengthy disclaimers that my yelling did not mean that I believed Roko was right about CEV-based agents [= Eliezer’s early model of indirectly normative agents that reason with ideal aggregated preferences] torturing people who had heard about Roko's idea. [...] What I considered to be obvious common sense was that you did not spread potential information hazards because it would be a crappy thing to do to someone. The problem wasn't Roko's post itself, about CEV, being correct.

I don't buy this explanation for EY actions. From his original comment, quoted in the wiki page:

"One might think that the possibility of CEV punishing people couldn't possibly be taken seriously enough by anyone to actually motivate them. But in fact one person at SIAI was severely worried by this, to the point of having terrible nightmares, though ve wishes to remain anonymous."

"YOU DO NOT THINK IN SUFFICIENT DETAIL ABOUT SUPERINTELLIGENCES CONSIDERING WHETHER OR NOT TO BLACKMAIL YOU. THAT IS THE ONLY POSSIBLE THING WHICH GIVES THEM A MOTIVE TO FOLLOW THROUGH ON THE BLACKMAIL. "

"... DO NOT THINK ABOUT DISTANT BLACKMAILERS in SUFFICIENT DETAIL that they have a motive toACTUALLY [sic] BLACKMAIL YOU. "

"Meanwhile I'm banning this post so that it doesn't (a) give people horrible nightmares and (b) give distant superintelligences a motive to follow through on blackmail against people dumb enough to think about them in sufficient detail, though, thankfully, I doubt anyone dumb enough to do this knows the sufficient detail. (I'm not sure I know the sufficient detail.) "

"You have to be really clever to come up with a genuinely dangerous thought. "

"... the gist of it was that he just did something that potentially gives superintelligences an increased motive to do extremely evil things in an attempt to blackmail us. It is the sort of thing you want to be EXTREMELY CONSERVATIVE about NOT DOING."

This is evidence that Yudkowsky believed, if not that Roko's argument was correct as it was, that at least it was plausible enough that could be developed in a correct argument, and he was genuinely scared by it.

It seems to me that Yudkowsky's position on the matter was unreasonable. LessWrong is a public forum unusually focused on discussion about AI safety, in particular at that time it was focused on discussion about decision theories and moral systems. What better place to discuss possible failure modes of an AI design?
If one takes AI risk seriously, and realized that an utilitarian/CEV/TDT/one-boxing/whatever AI might have a particularly catastrophic failure mode, the proper thing to do would be to publicly discuss it, so that the argument can be either refuted or accepted, and if it was accepted it would imply scrapping that particular AI design and making sure that anybody who may create an AI is aware of that failure mode. Yelling and trying to sweep it under the rug was irresponsible.

Comment author: RobbBB 06 October 2015 11:26:41AM *  8 points [-]

"One might think that the possibility of CEV punishing people couldn't possibly be taken seriously enough by anyone to actually motivate them. But in fact one person at SIAI was severely worried by this, to the point of having terrible nightmares, though ve wishes to remain anonymous."

This paragraph is not an Eliezer Yudkowsky quote; it's Eliezer quoting Roko. (The "ve" should be a tip-off.)

This is evidence that Yudkowsky believed, if not that Roko's argument was correct as it was, that at least it was plausible enough that could be developed in [sic] a correct argument, and he was genuinely scared by it.

If you kept going with your initial Eliezer quote, you'd have gotten to Eliezer himself saying he was worried a blackmail-type argument might work, though he didn't think Roko's original formulation worked:

"Again, I deleted that post not because I had decided that this thing probably presented a real hazard, but because I was afraid some unknown variant of it might, and because it seemed to me like the obvious General Procedure For Handling Things That Might Be Infohazards said you shouldn't post them to the Internet."

According to Eliezer, he had three separate reasons for the original ban: (1) he didn't want any additional people (beyond the one Roko cited) to obsess over the idea and get nightmares; (2) he was worried there might be some variant on Roko's argument that worked, and he wanted more formal assurances that this wasn't the case; and (3) he was just outraged at Roko. (Including outraged at him for doing something Roko thought would put people at risk of torture.)

What better place to discuss possible failure modes of an AI design? [...] Yelling and trying to sweep it under the rug was irresponsible.

There are lots of good reasons Eliezer shouldn't have banned R̶o̶k̶o̶ discussion of the basilisk, but I don't think this is one of them. If the basilisk was a real concern, that would imply that talking about it put people at risk of torture, so this is an obvious example of a topic you initially discuss in private channels and not on public websites. At the same time, if the basilisk wasn't risky to publicly discuss, then that also implies that it was a transparently bad argument and therefore not important to discuss. (Though it might be fine to discuss it for fun.)

Roko's original argument, though, could have been stated in one sentence: 'Utilitarianism implies you'll be willing to commit atrocities for the greater good; CEV is utilitarian; therefore CEV is immoral and dangerous.' At least, that's the version of the argument that has any bearing on the conclusion 'CEV has unacceptable moral consequences'. The other arguments are a distraction: 'utilitarianism means you'll accept arbitrarily atrocious tradeoffs' is a premise of Roko's argument rather than a conclusion, and 'CEV is utilitarian in the relevant sense' is likewise a premise. A more substantive discussion would have explicitly hashed out (a) whether SIAI/MIRI people wanted to construct a Roko-style utilitarian, and (b) whether this looks like one of those philosophical puzzles that needs to be solved by AI programmers vs. one that we can safely punt if we resolve other value learning problems.

I think we agree that's a useful debate topic, and we agree Eliezer's moderation action was dumb. However, I don't think we should reflexively publish 100% of the risky-looking information we think of so we can debate everything as publicly as possible. ('Publish everything risky' and 'ban others whenever they publish something risky' aren't the only two options.) Do we disagree about that?

Comment author: anon85 06 October 2015 04:22:26AM 2 points [-]

I think saying "Roko's arguments [...] weren't generally accepted by other Less Wrong users" is not giving the whole story. Yes, it is true that essentially nobody accepts Roko's arguments exactly as presented. But a lot of LW users at least thought something along these lines was plausible. Eliezer thought it was so plausible that he banned discussion of it (instead of saying "obviously, information hazards cannot exist in real life, so there is no danger discussing them").

In other words, while it is true that LWers didn't believe Roko's basilisk, they thought is was plausible instead of ridiculous. When people mock LW or Eliezer for believing in Roko's Basilisk, they are mistaken, but not completely mistaken - if they simply switched to mocking LW for believing the basilisk is plausible, they would be correct (though the mocking would still be mean, of course).

Comment author: RobbBB 06 October 2015 07:18:52AM *  1 point [-]

The wiki article talks more about this; I don't think I can give the whole story in a short, accessible way.

It's true that LessWrongers endorse ideas like AI catastrophe, Hofstadter's superrationality, one-boxing in Newcomb's problem, and various ideas in the neighborhood of utilitarianism; and those ideas are weird and controversial; and some criticism of Roko's basilisk are proxies for a criticism of one of those views. But in most cases it's a proxy for a criticism like 'LW users are panicky about weird obscure ideas in decision theory' (as in Auerbach's piece), 'LWers buy into Pascal's Wager', or 'LWers use Roko's Basilisk to scare up donations/support'.

So, yes, I think people's real criticisms aren't the same as their surface criticisms; but the real criticisms are at least as bad as the surface criticism, even from the perspective of someone who thinks LW users are wrong about AI, decision theory, meta-ethics, etc. For example, someone who thinks LWers are overly panicky about AI and overly fixated on decision theory should still reject Auerbach's assumption that LWers are irrationally panicky about Newcomb's Problem or acausal blackmail; the one doesn't follow from the other.

Comment author: hg00 19 September 2015 10:46:18PM *  0 points [-]

Off topic, but... since you are allegedly a strong feminist, maybe you can recommend specific feminist writing that you think is good reading as a counterpoint to SSC's stuff? I've read a fair amount of feminist writing but I haven't managed to find much that is as well-argued/thoughtful/reasonable as the SSC stuff. Ideally it would be feminist writing that attempts to respond to SSC directly.

Comment author: RobbBB 20 September 2015 08:08:23PM *  2 points [-]

"As well-argued/thoughtful/reasonable as the SSC stuff" is a pretty hard target to hit on most topics. Could you be more specific than "feminist writing", e.g., talk about a certain claim SSC makes that you'd like to see assessed by others?

Popehat's Shirts and Shirtiness strikes me as an especially high-quality post that's SSC-ish while coming out in favor of some ideas you see in feminism. Agenty Duck's Hasty Genderalizations gives a broader argument for worrying about gender bias. Ben Kuhn's On Inclusivity in Less Wrong is an example of a good feminism-relevant response to a specific SSC argument (though I don't necessarily agree 100% with Ben's arguments, and some of them may be dated at this point).

I'm not sure which of these is closest to what you're looking for. I find blogs like Thing of Things, Gruntled and Hinged, and The Unit of Caring useful both for clearly articulating ideas in the feminist memespace and for providing independent confirmation for some SSC views.

Comment author: [deleted] 17 September 2015 04:09:26AM 1 point [-]

I think it matters in this context. If these people are contrarian simply because they happen to have lots of different views, then it's irrelevant that they're contrarian. If they're contrarian because they're DRAWN towards contrarian views, it means they're biased towards cryonics.

In response to comment by [deleted] on Open thread, Sep. 14 - Sep. 20, 2015
Comment author: RobbBB 17 September 2015 09:31:45AM 0 points [-]

I agree it matters in this case, but it doesn't matter whether we use the word "contrarianism" vs. tabooing it.

Also, your summary assumes one of the points under dispute: whether it's possible to be good at arriving at true non-mainstream beliefs ('correct contrarianism'), or whether people who repeatedly outperform the mainstream are just lucky. 'Incorrect contrarianism' and 'correct-by-coincidence contrarianism' aren't the only two possibilities.

Comment author: [deleted] 17 September 2015 12:40:25AM *  1 point [-]

Then it might be that futurism is irrelevant, rather than being expertise-like or bias-like. (Unless we think 'studying X while lacking tight, reliable feedback loops' in this context is worse than 'neither studying X nor having tight, reliable feedback loops.')

In this case futurism is two things in these people: 1. A belief in expertise about the future. 2. A tendency towards optimism about the future. Combined, these mean that these people both think cryonics will work in the future, and are more confident in this assertion than warranted.

Thiel, Yudkowsky, Hanson, etc. use "contrarian" to mean someone who disagrees with mainstream views.

I don't think so... it's more someone who has the tendency(in the sense of an aesthetic preference) to disagree with mainstream views. In this case, they would tend to be drawn towards cryonics because it's out of the mainstream, which should give us less confidence that they're drawn towards cryonics because it's correct.

In response to comment by [deleted] on Open thread, Sep. 14 - Sep. 20, 2015
Comment author: RobbBB 17 September 2015 01:02:33AM *  0 points [-]

One of the most common ways they use the word "contrarian" is to refer to beliefs that are rejected by the mainstream, for whatever reason; by extension, contrarian people are people who hold contrarian beliefs. (E.g., Galileo is a standard example of a "correct contrarian" whether his primary motivation was rebelling against the establishment or discovering truth.) "Aesthetic preference" contrarianism is a separate idea; I don't think it matters which definition we use for "contrarianism".

Comment author: [deleted] 16 September 2015 08:20:43PM 2 points [-]

My typical heuristic for reliable experts (taken from Thinking Fast and Slow I think) is that if experts have tight, reliable feedback loops, they tend to be more trustworthy. Futurism obviously fails this test. Contrarianism isn't really a "field" in itself, and I tend to think of it more as a bias... although EY would obviously disagree.

In response to comment by [deleted] on Open thread, Sep. 14 - Sep. 20, 2015
Comment author: RobbBB 16 September 2015 11:09:59PM *  1 point [-]

My typical heuristic for reliable experts (taken from Thinking Fast and Slow I think) is that if experts have tight, reliable feedback loops, they tend to be more trustworthy. Futurism obviously fails this test.

Then it might be that futurism is irrelevant, rather than being expertise-like or bias-like. (Unless we think 'studying X while lacking tight, reliable feedback loops' in this context is worse than 'neither studying X nor having tight, reliable feedback loops.')

Contrarianism isn't really a "field" in itself, and I tend to think of it more as a bias...

Thiel, Yudkowsky, Hanson, etc. use "contrarian" to mean someone who disagrees with mainstream views. Most contrarians are wrong, though correct contrarians are more impressive than correct conformists (because it's harder to be right about topics where the mainstream is wrong).

Comment author: Lumifer 16 September 2015 04:18:36AM 1 point [-]

I don't think the average matters, it's the right tail of the distribution that's important.

Take, say, people with 130+ IQ -- that's about 2.5% of your standard white population and the overwhelming majority of them are not signed up. In fact, in any IQ quantile only a miniscule fraction has signed up.

Comment author: RobbBB 16 September 2015 07:37:32PM *  0 points [-]

entirelyuseless made the point that low cryonics use rates in the general population are evidence against the effectiveness of cryonics. James Miller responded by citing evidence supporting cryonics: that cryonicists are disproportionately intelligent/capable/well-informed. If your response to James is just that very few people have signed up for cryonics, then that's restating entirelyuseless' point. "The intellectual quality of some people who have NOT signed up for cryonics is exceptionally high" would be true even in a world where every cryonicist were more intelligent than every non-cryonicist, just given how few cryonicists there are.

View more: Prev | Next