ChristianKl comments on A big Singularity-themed Hollywood movie out in April offers many opportunities to talk about AI risk - Less Wrong Discussion
You are viewing a comment permalink. View the original post to see all comments and the full post content.
Loading…
Subscribe to RSS Feed
= f037147d6e6c911a85753b9abdedda8d)
Comments (84)
I agree that this a question that deserve serious thought. But the issue of violating the WIkipedia policy doens't factor much into the calculation.
It's quite natural behavior to add relevant quotations to a Wikipedia article. I wouldn't do it with an account without prior benign history or through anonymous edits.
If you are a good citizen of the web, you probably do fix Wikipedia errors when you notice them, so you should have an account that doesn't look spammy. If you don't, then you probably leave the task to someone else who has a better grasp on Wikipedia.
Good thing you're not discussing it in a public forum, then, where screencaps are possible.
The fact that the issue violates Wikipedia policy is an essential part of why doing as you propose would be likely to have a negative impact on MIRI's reputation.
(For the avoidance of doubt, I don't think this is the only reason not to do it. If you use something that has policies, you should generally follow those policies unless they're very unreasonable. But since ChristianKI is arguing that an expected-utility calculation produces results that swamp that (by tweaking the probability of a good/bad singularity) I think it's important to note that expected utility maximizing doesn't by any means obviously produce the conclusions he's arguing for.)