Comment author: handoflixue 25 December 2012 12:05:56AM 1 point [-]

Which category is it if you decide based on multiple factors, ONE of which is PR? And why is this a bad thing, if that's what you believe?

Comment author: kodos96 25 December 2012 12:43:53AM 1 point [-]

Before I spend any more time replying to this, can you clarify for me... do you and I actually disagree about something of substance here? I.e. how an organization should, in the real world, deal with PR concerns? Or are we just arguing about the most technically correct way to go about stating our position?

Comment author: handoflixue 25 December 2012 12:00:50AM *  0 points [-]

You're flat-out wrong about #3. Encryption is just a mathematical algorithm, it doesn't care who uses it, only that you have the key.

In short, encryption is just a very complex function, so you feed Key + Message in, and you get an Output. f(K,M) = O

I already have access to Key and Message, so I can share both of those. The only thing you can possibly secure is f().

If you have a cryptographic program, like OTR, I can just decompile it and get f(), and then post a modified version that lets the user manually configure their key (I think this is actually trivial in OTR, but it's been years since I poked at it)

If it's a website where I login and it auto-decrypts things for me, then I can just send someone the URL and the key I use.

Point 2 seems to reply on Point 3, and as far as I'm aware the only formally mathematically provably deniable method WHEN THE KEY IS COMPROMISED is a one-time pad.

I'm not sure how much crypto experience you have, but "and no one else knows the key" is a foundation of every algorithm I have ever worked on, and I'm reasonably confident that it's a mathematical requirement. I simply cannot imagine how you could possibly write a crypto algorithm that is secure EVEN with a compromised key.

EDIT: If you still think I'm wrong, can you please give me a sense of your crypto experience? For reference: I've met with the people who wrote OTR and hang out in a number of crypto circles, but only do fairly basic stuff in my actual work. I do still have a hobby interest in it, and follow it, but the last time I did any serious code breaking was about a decade ago.

Comment author: kodos96 25 December 2012 12:36:23AM *  -1 points [-]

You seem to be using a very narrow definition of "crypto".. I'm not sure whether you're just being pedantic about definitions, in which case you may be correct, or if you're actually disputing the substance of what I'm saying. To answer your question, I'm not a cryptographer, but I have a CS degree and am quite capable of reading and understanding crypto papers (though not of retaining the knowledge for long)... it's been several years since I read the relevant papers, so I might be getting some of the details wrong in how I'm explaining it, but the basic concept of deniable message authentication is something that's well understood by mainstream cryptographers.

You seem to be aware of the existence of OTR, so I'm confused - are you claiming that it doesn't accomplish what it says it does? Or just that something about the way I'm proposing to apply similar technology to this use case would break some of its assumptions? The latter case is entirely possible, as so far I've put a grand total of about 5 minutes thought into it... if that's the case I'd be curious to know what are the relevant assumptions my proposed use case would break?

Comment author: Eliezer_Yudkowsky 24 December 2012 11:12:05PM 2 points [-]

This seems to me like a deliberate misunderstanding. But taking it at face value, a story in which violence is committed against targets not analogous to any present-day identifiable people, or which is not committed for any reasons obviously analogous to present-day motives, is fine. The Sword of Good is not advocating for killing wizards who kill orcs, although Dolf does get his head cut off. Betrayed-spouse murder mysteries are not advocating killing adulterers - though it would be different if you named the victim after a specific celebrity and depicted the killer in a sympathetic light. As much as people who don't like this policy, might wish that it were impossible for anyone to tell the difference so that they could thereby argue against the policy, it's not actually very hard to tell the difference.

Comment author: kodos96 24 December 2012 11:23:38PM 2 points [-]

As much as people who don't like this policy, might wish that it were impossible for anyone to tell the difference so that they could thereby argue against the policy, it's not actually very hard to tell the difference.

I didn't interpret CronoDAS's post as intending to actually advocate violence. I viewed it as really silly and kind of dickish, and a good thing that he ultimately removed it, but an actual call to violence? No. It was a thought experiment. His thought experiment was set in the present day, while yours was set in the far future, but other than that I don't see a bright line separating them.

It may not be be very hard for you to tell the difference, since you wrote the policy, so you may very well have a clear bright line separating the two in your head, but we don't.

Comment author: kodos96 24 December 2012 10:44:37PM *  11 points [-]

I am asking in advance if anyone has non-obvious consequences they want to point out or policy considerations they would like to raise. In other words, the form of this discussion is not 'Do you like this?' - you probably have a different cost function from people who are held responsible for how LW looks as a whole - but rather, 'Are there any predictable consequences we didn't think of that you would like to point out

Eliezer, at this point I think it's fair to ask: has anything anyone has said so far caused you to update? If not, why not?

I realize some of my replies to you in this thread have been rather harsh, so perhaps I should take this opportunity to clarify: I consider myself a big fan of yours. I think you're a brilliant guy, and I agree with you on just about everything regarding FAI, x-risk, SIAI's mission.... I think you're probably mankind's best bet if we want to successfully navigate the singularity. But at the same time, I also think you can demonstrate some remarkably poor judgement from time to time... hey, we're all running on corrupted hardware after all. It's the combination of these two facts that really bothers me.

I don't know of any way to say this that isn't going to come off sounding horribly condescending, so I'm just going to say it, and hope you evaluate it in the context of the fact that I'm a big fan of your work, and in the grand scheme of things, we're on the same side.

I think what's going on here is that your feelings have gotten hurt by various people misattributing various positions to you that you don't actually hold. That's totally understandable. But I think you're confusing the extent to which your feelings have been hurt with the extent to which actual harm has been done to SIAI's mission, and are overreacting as a result. I'm not a psychologist - this is just armchair speculation.... I'm just telling you how it looks from the outside.

Again, we're all running on corrupted hardware, so it's entirely natural for even the best amongst us to make these kinds of mistakes... I don't expect you to be an emotionless Straw Vulcan (and indeed, I wouldn't trust you if you were)... but your apparent unwillingness to update in response to other's input when it comes to certain emotionally charged issues is very troubling to me.

So to answer your question "Are there any predictable consequences we didn't think of that you would like to point out"... well I've pointed out many already, but the most concise, and most important predictable consequence of this policy which I believe you're failing to take into account, is this: IT LOOKS HORRIBLE... like, really really bad. Way worse than the things it's intended to combat.

Comment author: handoflixue 24 December 2012 09:51:17PM 2 points [-]

Implement a full on OTR style system providing full deniability through crypto. Rather than stopping content from being copied, just make sure you can claim any copy is a forgery, and nobody can prove you wrong. A MAJOR engineering effort of course, but totally possible, and 100% effective.

I can't help but see two major flaws:

1) If I link to a major, encrypted offshoot of LessWrong, people will AUTOMATICALLY be suspicious and it will damage PR.

2) Why would it be any easier to cry "it's a forgery" in this situation versus me posting a screenshot of an unencrypted forum? o.o Especially given #1...

3) I can share my password / decryption key / etc..

Comment author: kodos96 24 December 2012 09:58:55PM *  -1 points [-]

Well, point 3 can be eliminated by proper use of crypto. See OTR

The response to point 2 is that by having it be publicly known to everyone that messages' contents are formally mathematically provably deniable (as can be guaranteed by proper crypto implementation), that disincentivizes people from even bothering to re-post content in the first place.

Point 1, however, I agree with completely, and that's why I'm not actually advocating this solution.

Comment author: handoflixue 24 December 2012 09:29:26PM -1 points [-]

Taboo "driving" and "informing" and explain the difference between those two to me?

Or we can save ourselves some time if this resolves your objection: Eliezer is saying that he is adding the OPTION to censor things if they are a PR problem OR because the person is needlessly incriminating themselves. I'm not sure how that's a bad OPTION to have, given that he's explicitly stated he will not mindlessly enforce it, and in fact has currently enforced it zero (0) times to my knowledge (the post that prompted this was voluntarily withdrawn by it's author)

Comment author: kodos96 24 December 2012 09:32:05PM 3 points [-]

One the one hand, you're deciding policy based on non-PR related factors, then thinking about the most PR friendly way to proceed from there. On the other hand, you're letting PR actually determine policy.

Comment author: handoflixue 24 December 2012 09:18:27PM 0 points [-]

"it might make us look bad" is a horrible argument

You can argue that LessWrong shouldn't care about PR, or that censorship is going to be bad PR, or that censorship is unnecessary, but you can't argue that PR is a fundamentally horrible idea without some very strong evidence (which you did not provide).

-

It's almost tautological that if a group cares about PR, it HAS to care about what makes them look bad:

If Obama went on record saying that we should kill everyone on Less Wrong, and made it clear he was serious, I'd hope to high hell that there would be an impeachment trial.

If Greenpeace said we should kill all the oil CEOs, people would consider them insane terrorists.

If the oil CEOs suggested that there might be... incentives... should Greenpeace members be killed...

Comment author: kodos96 24 December 2012 09:25:23PM 1 point [-]

You can argue that LessWrong shouldn't care about PR, or that censorship is going to be bad PR, or that censorship is unnecessary, but you can't argue that PR is a fundamentally horrible idea without some very strong evidence (which you did not provide).

That was perhaps a bit of an overstatement on my part. Considering PR consequences of actions is certainly a good thing to do. But if PR concerns are driving your policy, rather than simply informing it, that's bad.

Comment author: Eliezer_Yudkowsky 24 December 2012 07:51:55PM 1 point [-]

Well, are you?

No. To prove this, I shall shortly delete the post advocating it.

True, but you have said things that seem to imply it. Seriously, you can't go around saying "X" and "X->Y" and then object when people start attributing position "Y" to you.

Point one: We never said X->Y. We said X, and a bunch of people too stupid to understand the fallacy of appeal to consequences said 'X->violence, look what those bad people advocate' as an attempted counterargument. Since no actual good can possibly come of discussing this on any set of assumptions, it would be nice to have the counter-counterargument, "Unlike this bad person here, we have a policy of deleting posts which claim Q->specific-violence even if the post claims not to believe in Q because the identifiable target would have a reasonable complaint of being threatened".

Comment author: kodos96 24 December 2012 09:11:07PM 16 points [-]

it would be nice to have the counter-counterargument, "Unlike this bad person here, we have a policy of deleting posts which claim Q->specific-violence even if the post claims not to believe in Q because the identifiable target would have a reasonable complaint of being threatened".

I would find this counter-counter-argument extremely uncompelling if made by an opponent. Suppose you read someone's blog who made statements which could be interpreted as vaguely anti-Semitic, but it could go either way. Now suppose someone in the comments of that blog post replied by saying "Yeah, you're totally right, we should kill all the Jews!".

Which type of response from the blog owner do you think would be more likely to convince you that he was not actually an anti-Semite: 1) deleting the comment, covering up its existence, and never speaking of it, or 2) Leaving the comment in place, and refuting it - carefully laying out why the commenter is wrong.

I know that I for one would find the latter response much more convincing of the author's benign intent.

Note: in order to post this comment, despite it being, IMHO entirely on-point and important to the conversation, I had to take a 5 point karma hit.... due to the LAST poorly thought out, dictatorially imposed, consensus-defying policy change.

Comment author: roystgnr 24 December 2012 07:10:51PM 2 points [-]

This must be why the media companies haven't given up on DRM yet. They think if they can just unmask and arrest the ringleaders of the "organized conspiracy out there" then copy protection will start working, when in reality any random person can become a "conspiracy" member with nothing more than a little technical knowledge, a little free time, and a moral code that encourages copying.

To be fair, the "vetting" and "full deniability" options don't really apply to the ??AA. The best pre-existing example for those kinds of policies might be the Freemasons or the Mormons? In neither case would I be confident that the bad PR they've avoided by hiding embarrassing things hasn't been worse than the bad PR they've abetted by obviously dissembling and/or by increasing the suspicion that they're hiding even worse things.

Comment author: kodos96 24 December 2012 08:14:06PM *  0 points [-]

In neither case would I be confident that the bad PR they've avoided by hiding embarrassing things hasn't been worse than the bad PR they've abetted by obviously dissembling and/or by increasing the suspicion that they're hiding even worse things.

Exactly. That's why I'm not actually advocating any of these technical solutions, just pointing out that they do exist in solution-space.

The solution that I'm actually advocating is even simpler still: do nothing. Rely on self-policing and the "don't be an asshole" principle, and in the event that that fails (which it hasn't yet), then counter bad speech with more speech: clearly state "LW/SIAI does not endorse this suggestion, and renounces the use of violence." If people out there still insist on slandering SIAI by association to something some random guy on LW said, then fuck em - haters gonna hate.

Comment author: Qiaochu_Yuan 24 December 2012 11:13:53AM *  2 points [-]

I wouldn't have posted the following except that I share Esar's concerns about representativeness:

I think this is a good idea. I think using the word "censorship" primes a large segment of the LW population in an unproductive direction. I think various people are interpreting "may be deleted" to mean "must be deleted." I think various people are blithely ignoring this part of the OP (emphasis added):

In other words, the form of this discussion is not 'Do you like this?' - you probably have a different cost function from people who are held responsible for how LW looks as a whole

In particular, I think people are underestimating how important it is for LW not to look too bad, and also underestimating how bad LW could be made to look by discussions of the type under consideration.

Finally, I strongly agree that

anyone talking about a proposed crime on the Internet fails forever as a criminal[.]

Comment author: kodos96 24 December 2012 05:34:33PM 0 points [-]

In particular, I think people are underestimating how important it is for LW not to look too bad

I'm not underestimating that at all... I'm saying that this policy makes us look bad... WAY worse than the disease it's intended to cure, especially in light of the fact that that disease cleared itself up in a few hours with no intervention necessary.

View more: Prev | Next