Comment author: ChristianKl 15 October 2013 11:41:58AM 2 points [-]

Did you play against AI that do have won sometime in the past?

Comment author: Sly 16 October 2013 02:40:58AM 0 points [-]

I do not honestly know. I will happily play a "hard" opponent like Eliezer or Tux. I have said this before, I estimate 99%+ chance of victory.

Comment author: falenas108 14 October 2013 09:28:31PM 8 points [-]

Other people have expressed similar sentiments, and then played the AI Box experiment. Even of the ones who didn't lose, they still updated to "definitely could have lost in a similar scenario."

Unless you have reason to believe your skepticism comes from a different place than theirs, you should update towards gatekeeping being harder than you think.

Comment author: Sly 15 October 2013 03:04:35AM *  4 points [-]

I have played the game twice and updated in the opposite direction you claim.

In fact, my victories were rather trivial. This is despite the AIs trying really really hard.

Comment author: Athrelon 06 October 2013 03:41:52PM 0 points [-]

This is true; however keeping a website running is still very, very cheap compared to almost anything else the government does, including functions that are continuing as usual during the shutdown.

If web apps are too high maintenance, that does not explain the shutdown of government Twitters (example: https://twitter.com/NOAA, which went to the extra effort of posting that "we won't be tweeting 'cause shutdown.") I note with amusement however that the Health and Human Services Twitter is alive and well and tweeting about the ACA.

Comment author: Sly 06 October 2013 06:36:35PM 0 points [-]

"This is true; however keeping a website running is still very, very cheap compared to almost anything else the government does, including functions that are continuing as usual during the shutdown."

This is literally irrelevant when the non-essential services have to be shut down. If your techs get furloughed, shutting down the site is appropriate.

The twitter accounts are "shut down" in the sense that the employee who would have done the tweeting is now furloughed and can't. Putting out a tweet explaining the upcoming lapse makes a whole lot of sense to me.

Comment author: [deleted] 02 October 2013 06:15:45AM *  10 points [-]

Interesting statements I ran into with regards to kabuki theater aspects of the so called United States federal government shutdown of 2013. This resulted in among other things closing down websites.

A website shouldn't just go down when the people managing it stop working, it's not like they're pedaling away inside the servers. Block the federal highways with army tanks, sorry the government is closed.

There is a nontrivial set of the voting public who legitimately believe money equals tech working via magical alchemy.

I was interested to know this kind of thing has a name: Washington Monument Syndrome.

The name derives from the National Park Service's alleged habit of saying that any cuts would lead to an immediate closure of the wildly popular Washington Monument.

Comment author: Sly 02 October 2013 09:02:17PM 8 points [-]

This is actually a terrible example of Washington Monument Syndrome.

" Hi, Server admin here... We cost money as does our infrastructure, I imagine a site that large costs a very good deal, we aren't talking five bucks on bluehost here.

I am private sector, but if I were to be furloughed for an indeterminate amount of time you really have two options. Leave things on autopilot until the servers inevitably break or the site crashes at which point parts or all of it will be left broken without notice or explanation. Or put up a splash page and spin down 99% of my infrastructure (That splash page can run on a five dollar bluehost account) and then leave. I won't be able to come in while furloughed to put it up after it crashes.

If you really think web apps keep themselves running 24/7 without intervention we really have been doing a great job with that illusion and I guess the sleepless nights have been worth it to be successfully taken for-granted."

Comment author: Ishaan 30 September 2013 03:38:23AM 1 point [-]

Possibly, but what about the descriptions of emotional turmoil? I'm assuming the report of the game isn't all part of the role-play.

Comment author: Sly 01 October 2013 12:01:36AM 0 points [-]

I know that I personally go into competitive games with a different mindset than the mindset I have when roleplaying.

If they went into it trying to roleplay emotions should be expected. Reporting that turmoil in the report is just accurate reporting.

Comment author: Ishaan 29 September 2013 07:04:35AM *  2 points [-]

Yeah, winning is trivial - you just don't open the damn box. It can't get more trivial than that. (Although, you didn't say whether or not your opponent had proved themselves by winning as AI against others a few times?)

It's still worth thinking about though, because something about my model of humans is off.

I didn't expect so many people to lose. I just don't know how to update my model of people to one where there are so many people who could lose the AI box game. The only other major thing I can think of that persists to challenge my model in this way (and continues to invite my skepticism despite seemingly trustworthy sources) is hypnosis.

It's possible the two have common root and I can explain two observations with one update.

Comment author: Sly 29 September 2013 10:41:56PM 0 points [-]

I think a lot of gatekeepers go into it not actually wanting to win. If you go in just trying to have fun and trying to roleplay, that is different than trying to win a game.

Comment author: Ishaan 28 September 2013 05:39:27AM *  10 points [-]

Assuming none of this is fabricated or exaggerated, every time I read these I feel like something is really wrong with my imagination. I can sort of imagine someone agreeing to let the AI out of the box, but I fully admit that I can't really imagine anything that would elicit these sorts of emotions between two mentally healthy parties communicating by text-only terminals, especially with the prohibition on real-world consequences. I also can't imagine what sort of unethical actions could be committed within these bounds, given the explicitly worded consent form. Even if you knew a lot of things about me personally, as long as you weren't allowed to actually, real-world, blackmail me...I just can't see these intense emotional exchanges happening.

Am I the only one here? Am I just not imagining hard enough? I'm actually at the point where I'm leaning towards the whole thing being fabricated - fiction is more confusing than truth, etc. If it isn't fabricated, I hope that statement is taken not as an accusation, but as an expression of how strange this whole thing seems to me, that my incredulity is straining through despite the incredible extent to which the people making claims seem trustworthy.

Comment author: Sly 29 September 2013 06:16:37AM 0 points [-]

You are correct here. The only keepers losing are people who do not actually know how to win.

I have played twice, and victory was trivial.

Comment author: Risto_Saarelma 16 July 2013 12:23:17PM 5 points [-]

Ben Goertzel will take your money and try put an AGI inside a robot.

Trigger warning: Those creepy semi-human robots that will make anyone who hasn't spent months and months locked in a workshop making them do those human-imitating jerky facial gestures recoil in horror.

Comment author: Sly 17 July 2013 08:52:26AM 1 point [-]

That was hideous. Poor production values and a sloppy video that oozes incompetence.

In response to comment by Sly on Why one-box?
Comment author: Decius 10 July 2013 07:03:06PM 1 point [-]

Suppose my strategy made an equal (enough) number of suggestions for each option over the last 1m trials, while the opponent played paper every time. My current strategy suggests that playing rock on the next game is the best move. The opponent's move is defined to not be dependent on my prior moves (because otherwise things get too complicated for brief analysis)

There are two major competing posterior strategies at this point: "Scissors for the first 1M trials, then rock" and "Scissors for the first 1M trials" It is not possible for my prior probability for "Scissors for the first N, then rock" to be higher than my probability for "Scissors forever" for an infinite number of N, so there is some number of trials after which any legal prior probability distribution favors "Scissors forever", if it loses only a finite number of times.

At this point I'm going to try to save face by pointing out that for each N, there is a legal set of prior probabilities of the optimum strategy to suggest each option an equal number of time. They would have to be arranged such that "My opponent will play paper X times then something else" is more likely than "My opponent will play paper X times then play paper again" for 2/3 of X from 0 to N. Given that "My opponent will always play paper" is a superset of the latter, and each time I am wrong I must eliminate a probability space larger than it from consideration, and that I have been wrong 700k times, I obviously must have assigned less than ~1e-6 initial probability to all estimates that my opponent will play paper 1M+1 times in a row, but higher than that to ~700k cases of supersets of "my opponent will play paper X times in a row then change" where X is less than 1M. While a legal set of priors, I think it would be clearly unreasonable in practice to fail to adapt to a strategy of strict paper within 10.

Strangely, many of the strategies which are effective against humans for best-of-seven seem to be ineffective against rational agents for long-term performance. Would it be interesting to have a RPS best-of competition between programs with and without access to their opponent's source code, or even just between LW readers who are willing to play high-stakes RPS?

In response to comment by Decius on Why one-box?
Comment author: Sly 13 July 2013 07:52:14PM 1 point [-]

Cool, sounds like we are converging.

I would be interested in seeing a RPS competition between programs, sounds interesting.

In response to comment by Sly on Why one-box?
Comment author: Decius 09 July 2013 12:23:00PM 0 points [-]

Suppose your opponent has thrown paper N (or X%) times and won every time they did. Is that evidence for, or evidence against, the proposition that they will play paper in the next trial? (or does the direction of evidence vary with N or X?)

In response to comment by Decius on Why one-box?
Comment author: Sly 10 July 2013 08:08:34AM 0 points [-]

"Suppose your opponent has thrown paper N (or X%) times and won every time they did. Is that evidence for, or evidence against, the proposition that they will play paper in the next trial? (or does the direction of evidence vary with N or X?)"

All of this is irrelevant.

So I will admit I am frustrated here. I don't think that your analogy is even close to equivalent,

I think you are thinking about this in the wrong way.

So let's say you were an adviser advising one of the players on what to choose. Every time you told him to throw rock over the last million games, he lost. Yet every time you told him to throw Scissors he won. Now you have thought very much about this problem, and all of your theorizing keeps telling you that your player should play Rock (the theorycrafting has told you this for quite a while now).

At what point is this evidence that you are reasoning incorrectly about the problem, and really you should just tell the player to play scissors? Would you actually continue to tell him to throw Rock if you were losing $1 every time the player you advised lost?

Now if this advising situation had been a game that you played with your strategy and I had separately played with my strategy, who would have won?

View more: Prev | Next