Luke_A_Somers comments on I played as a Gatekeeper and came pretty close to losing in a couple of occasions. Logs and a brief recap inside. - Less Wrong

5 [deleted] 08 February 2015 04:32PM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (46)

You are viewing a single comment's thread.

Comment author: Luke_A_Somers 08 February 2015 04:53:21PM 7 points [-]

Whoa, someone actually letting the transcript out. Has that ever been done before?

Comment author: [deleted] 08 February 2015 04:59:19PM 8 points [-]

Actually, it has been done several times, but most of them are pretty boring.

Comment author: shminux 08 February 2015 05:47:42PM 4 points [-]

I still don't recall any where the gatekeeper lost.

Comment author: habeuscuppus 11 February 2015 09:36:32PM 0 points [-]

In general it seems that gatekeepers who win are more willing to release the transcripts.

It's also possible that the 'best' AI players are the ones most willing to pre-commit to not releasing transcripts, as not having your decisions (or the discussions that led to them) go public helps eliminate that particular disincentive to releasing the AI from the box.

Comment author: lmm 13 February 2015 08:08:43PM 0 points [-]

Never still seems extraordinary. I find myself entertaining hypotheses like "maybe the AI has never actually won".

Comment author: habeuscuppus 16 February 2015 06:02:06PM 0 points [-]

Eliezer Yudkowsky has been let out as the AI at least twice[1][2] but both tests were precommitted to secrecy.

I'd be surprised if he's the only one who has ever won as the AI, I think it more likely that this is a visibility issue (e.g. despite him being a very-high profile person in the AI safety memetic culture, you weren't aware that Eliezer had won as the AI when you made your comment) and while I'm not aware of others who have won as the AI, I would place my bet on that being merely a lack of knowledge on my part, and not because no one else actually has.

this is further compounded by the fact that some (many?) games are conducted under a pre-commitment to secrecy, and the results that get the most discussion (and therefore, most visibility) are the ones with full transcripts for third-parties to pick through.

Comment author: lmm 16 February 2015 07:53:45PM *  0 points [-]

I was already aware of those public statements. I remain rather less than perfectly confident that Yudkowsky actually won.

Comment author: habeuscuppus 16 February 2015 07:59:29PM 0 points [-]

forgive me if I misunderstand you, but you seem to be implying that, on two separate occasions, two different people were (induced to?) lie about the outcome of an experiment.

So you're implying that either Eliezer is dishonest, or both of his opponents were dishonest on his behalf. And you find this more likely than an actual AI win in the game?

Comment author: Luke_A_Somers 09 February 2015 02:39:54AM 1 point [-]

Awww. I didn't actually read this one either, yet. Is this one boring?

Comment author: MathiasZaman 09 February 2015 01:40:04PM 0 points [-]

I didn't found it particularly interesting. Entertaining the idea of letting the AI out is far from the same as almost letting the AI out.

Comment author: [deleted] 09 February 2015 11:58:01AM *  0 points [-]

I can't speak for myself, but at least it wasn't boring to play. Polymathwannabe also said that he enjoyed the experiment enormously.

Comment author: gjm 09 February 2015 12:39:36AM 5 points [-]

Did you deliberately phrase that ("letting the transcript out") so as to hint at an AI-Box-Box game, in which one player's goal is to convince the other to release the transcript of an earlier AI-Box game, while the other tries to keep it secret?

Comment author: Luke_A_Somers 09 February 2015 02:39:22AM 0 points [-]

I probably had the phrasing primed and ready to go in my brain, but it wasn't intentional.

Comment author: RedErin 10 February 2015 08:08:00PM 1 point [-]

Whoa, someone actually letting the transcript out. Has that ever been done before?

Yes, but only when the gatekeeper wins. If the AI wins, then they wouldn't want the transcript to get out, because then their strategy would be less effective next time they played.

Comment author: Jiro 17 February 2015 05:16:18PM 0 points [-]

I would imagine that if we ever actually build such an AI, we would conduct some AI-box experiments to determine some AI strategies and figure out how to counter them. Humans who become the gatekeeper for the actual AI would be given the transcripts of AI-box experiment sessions to study as part of their gatekeeper training.

Letting out the transcript, then, would be a good thing. It would make the AI player's job harder because in the next experiment the human player will be aware of those strategies, but when facing an actual AI, the human will be aware of those strategies.

Comment author: lmm 13 February 2015 07:24:30PM 0 points [-]

Doesn't the same logic apply to the gatekeeper?

Comment author: RedErin 13 February 2015 09:20:33PM 0 points [-]

The Gatekeeper usually wants to publish if they win, to brag. Their strategy isn't usually a secret, it's simply to resist.