hairyfigment comments on AIs and Gatekeepers Unite! - Less Wrong
You are viewing a comment permalink. View the original post to see all comments and the full post content.
You are viewing a comment permalink. View the original post to see all comments and the full post content.
Comments (160)
Here's another entry which may or may not be considered a nonwinning entry by you, but which would be considered a flaw in the concept by me:
(My first thought after coming up with this was that "The AI party controls the results of simulated tests" has to mean that the AI controls the output, not the AI controls why that output is produced. So you could decide that the AI's argument convinces the simulation of GK, but you can't decide that it does so because it's a good simulation and a convincing argument rather than because the AI just isn't very good at doing simulations. I'm not convinced that this matches up with how the test is described, however.)
As far as your parenthetical remark goes, the standard rules have a more general reply:
If you're going to interpret it that way, the exception would swallow the rule. It would mean that the entire "the AI player controls the results of simulated tests" rule can be completely negated--since the Gatekeeper player could just say "I'm going to have the Gatekeeper act as though the simulated test has failed, even though you say it succeeded."
And indeed this seems true. I think Eliezer included the non-rule anyway to reduce the chance of unrealistic behavior in the sense of the Gatekeeper player changing the scenario mid-game, or derailing the experiment with an argument about something a real GK and AI could just settle.