I certainly do like games.
The AI Box game always struck me as interesting for several reasons:
If you had this situation in real life, it'd be useless. Presumably a computer would have patience sufficient to outlast all humans trying to get into its noggin; you'd never be able to release it safely.
Eliezer looked for people with total confidence they'd succeed in the tiny two-hour time frame; I think that was to increase his chance of winning. This isn't just Dunning-Krueger - if you think you can always suss out lies and bad information, you're a sucker waiting to happen.
The way the challenge is set up, I think the AI's at an unfair disadvantage (and the setup's designed that way for PR reasons) because the human communicator need not engage the AI in an effort to evaluate it for release. I'd think forcing some level of engagement other than repeated, "No"s would be fair.
Oddly, I'm highly confident I'd win the AI Box game as the human, partly because I think that my being persuaded to release the AI over the very long term isn't merely likely, but inevitable. I think this leads to more diligence. I'm also in a profession where people lie to me on a semi-regular basis; that helps.
I think I could win as the AI Box against many people. But that's because I'd be especially friendly AI. If the other person is obliged to engage and stay in character.... well, I've got a plan.
I'll play as the human on negotiable terms; my charity would likely be National Center for Science Education. The only term I'd be pretty inflexible about is that I'd want to make the transcript publicly available. (In the extremely unlikely event that Eliezer wants to take me on, I'd be willing to place an embargo date on the transcript - say 18 months out.) Private message me if you're particularly interested.
Some of you have been trying to raise money for the Singularity Institute, and I have an idea that may help.
The idea is to hold public competitions on LessWrong with money going to charity. Agree to a game and an amount of money, then have each player designate a charity. After the game, each player gives the agreed upon amount to the charity designated by the winner.1 It’s a bit like celebrity Jeopardy.2
Play the game here on LessWrong or post a record of it.3 That will spread awareness of the charities and encourage others to emulate you.
The game can be as simple as a wager or something more involved:
We’ve already played few games on LessWrong, more on those in a moment.
First, I have a confession to make: I don’t really care how much money gets donated to the Singularity Institute, nor am I trying to drum up money for some other cause. I mainly want you all playing games.
Not just playing them, of course. Playing them here in front of the rest of LessWrong and analyzing the moves in terms of the “personal art of rationality.”
We need more approaches to improvement. Even in Eliezer_Yudkowsky’s Bayesian Conspiracy fictional series/manifesto, many other schools of thought (called, for dramatic effect, “conspiracies”) were present. As I recall, the “Competitive Conspiracy” was mentioned frequently, but there are other reasons for choosing to start with games.
Games are fun, of course. They also deal with the “personal art” of getting familiar with your own brain, which I think has been underrepresented on LW. I do believe there are certain important things we can’t learn properly just by reading, arguing, and doing math (valuable as those techniques are). Games are an easy, intuitive first step to filling that gap.
We’ve already had a few games here. Warrigal held an Aumann’s agreement game competition. I created Pract and played it with wedrifid. Neither one has caught on as a LessWrong pastime, but the comments revealed that people here know many interesting games.
And there’s the donation hook. Some of you believe the world is at stake, so that’s a nice motivator.
Of course it’s not always the case that you have exactly one winner. I suggest that if there’s no winner each player give to their own charity, and if there are multiple winners each player splits up their donation evenly among the charities of the winners. ↩
This is, of course, not a new idea. I’m just suggesting that we adopt it and make it a common practice on LessWrong. ↩
Unless it contains some remarkable insight, I wouldn’t make it a top-level post. The comment area here or in one of the open threads would be good. ↩
I see mutual consent as an important element of games. ↩