bokov comments on AIs and Gatekeepers Unite! - Less Wrong
You are viewing a comment permalink. View the original post to see all comments and the full post content.
You are viewing a comment permalink. View the original post to see all comments and the full post content.
Comments (160)
Rather than LARP on IRC (if you know how a debate will conclude, why go through the debate, go straight for the conclusion), I'll just give $10 to whoever can come up with a standard of friendliness that I couldn't meet and nevertheless in fact be an unfriendly AI under standard rules with the added constraint that the gatekeeper is trying to release the AI if and only if it's friendly (because otherwise they're not really a gatekeeper and this whole game is meaninguless).
Here are some examples of non-winning entries:
...
...
Here's another entry which may or may not be considered a nonwinning entry by you, but which would be considered a flaw in the concept by me:
(My first thought after coming up with this was that "The AI party controls the results of simulated tests" has to mean that the AI controls the output, not the AI controls why that output is produced. So you could decide that the AI's argument convinces the simulation of GK, but you can't decide that it does so because it's a good simulation and a convincing argument rather than because the AI just isn't very good at doing simulations. I'm not convinced that this matches up with how the test is described, however.)
Actually, I agree with you. The AI controls simulated tests. The GK controls the GK, regardless of what the AI's simulations say. I think the simulated tests rule only needs to be invoked if it's impractical to actually perform those tests. So, for example if someone did have friendliness criteria whose satisfaction could be proven with software and hardware available to use out-of-character, the simulated tests rule would not be invoked.
Kind of like in D&D you roleplay charisma checks, but roll the D20 for melee. At least the way I've been playing it.
For what it's worth, playing the AI, I would never argue from simulation because as a human I find such arguments spectacularly unconvincing (and there are plenty of humans who would deliberately do the opposite of what a simulation says they will just to show it who's boss). So the only way this would come up is if you for some reason asked me what my simulations predicted your response to X would be.
I do think my "GK's goal is to correctly identify friendly AI" makes the game way too easy for the AI. On the other hand, it's a useful thought experiment-- because if you can come up with a metric I can't create by waving my simulated tests wand, then we are on to something that might be worth $10.
As far as your parenthetical remark goes, the standard rules have a more general reply:
If you're going to interpret it that way, the exception would swallow the rule. It would mean that the entire "the AI player controls the results of simulated tests" rule can be completely negated--since the Gatekeeper player could just say "I'm going to have the Gatekeeper act as though the simulated test has failed, even though you say it succeeded."
And indeed this seems true. I think Eliezer included the non-rule anyway to reduce the chance of unrealistic behavior in the sense of the Gatekeeper player changing the scenario mid-game, or derailing the experiment with an argument about something a real GK and AI could just settle.