You're looking at Less Wrong's discussion board. This includes all posts, including those that haven't been promoted to the front page yet. For more information, see About Less Wrong.

calef comments on I played the AI Box Experiment again! (and lost both games) - Less Wrong Discussion

35 Post author: Tuxedage 27 September 2013 02:32AM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (123)

You are viewing a single comment's thread.

Comment author: calef 27 September 2013 04:48:13AM 14 points [-]

The secrecy aspect of these games continues to rub me the wrong way.

I understand the argument--that an enumeration of strategies an oracle A.I. might take would only serve as a list of things a critic could point to and claim, "None of these would ever convince me!".

But the alternative is that critics continue to claim "an oracle A.I. could never convince me!", and the only 'critics' whose minds have been changed are actually just skeptical readers of lesswrong.com already familiar with the arguments of friendly A.I. who happen to invest multiple hours of time actually partaking in a simulation of the whole procedure.

So I suppose my point is two-fold:

  1. Anonymous testimony without chatlogs don't actually convince skeptics of anything.

  2. Discussions of actual strategies at worst inform readers of avenues of attack the readers might not have thought about, and at double worst supply people that probably won't ever be convinced that oracle AIs might be dangerous with a list of things to pretend they're immune to.

Comment author: chaosmage 30 September 2013 06:12:58PM *  4 points [-]

Even someone who isn't persuaded by an "AI" character in a log will come away with the impression that AIs could be particularly persuasive. In a world where most people don't really imagine AIs, this impression might be relevant news for a lot of people and can only help FAI research.

Comment author: RomeoStevens 27 September 2013 05:47:21AM 4 points [-]

Reading a log and engaging in a conversation are very different experiences.

Comment author: RobbBB 27 September 2013 08:39:53AM 9 points [-]

I'm not so sure we'd gain that much larger of an audience by peering under the hood. I'd expect the demystifying effect and hindsight bias to counteract most of the persuasive power of hard details, though I suppose only Eliezer, Tuxedage, and their guardians can determine that.

But I'm also concerned that this might drag our community a bit too far into AI-Box obsession. This should just be a cute thought experiment, not a blood sport; I don't want to see people get hurt by it unless we're especially confident that key minds will be changed. Some of the Dark Arts exhibited in these games are probably harmful to know about, and having the logs on the public Internet associated with LessWrong could look pretty awful. Again, this is something only the participants can determine.