SilasBarta comments on The flawed Turing test: language, understanding, and partial p-zombies - Less Wrong

11 Post author: Stuart_Armstrong 17 May 2013 02:02PM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (184)

You are viewing a single comment's thread.

Comment author: SilasBarta 17 May 2013 06:04:48PM *  6 points [-]

I don't think the "over-fitting" problem applies to the Turing Test: you can ask the candidate about anything, and adapt your later questions accordingly. There are proofs in computational complexity (that I'm too lazy to look up right now) that show that you can't pass this kind of test (except with exponentially small probability) but by containing a polynomial-time algorithm for the entire problem space. (It's related to the question of what problems are IP-complete -- i.e. the hardest among those problems that can be quickly solved via interactive proof.)

It would only be analogous to the test of the students if you published a short list of acceptable topics for the TT and limited the questions to that. Which they don't do.

Edit: If you were right, it would be much easier to construct such a "conversation savant" than it has proven to be.

Comment author: Stuart_Armstrong 17 May 2013 08:19:06PM 4 points [-]

If you were right, it would be much easier to construct such a "conversation savant" than it has proven to be.

Watson shocked me - I didn't think that type of performance was possible without AI completeness. That was a type of savant that I thought couldn't happen before AGI.

It might be that passing for a standard human in a Turing test is actually impossible without AGI - I'm just saying that I would want more proof in the optimised-for-Turing-test situation than in others.

Comment author: Vaniver 17 May 2013 08:39:49PM 5 points [-]

That was a type of savant that I thought couldn't happen before AGI.

This interests me (as someone professionally involved in the creation of savants, though not linguistic ones). Can you articulate why you thought that?

Comment author: Stuart_Armstrong 18 May 2013 06:19:26PM 1 point [-]

It wasn't formalised thinking. I bought into the idea of AI-complete problems, ie that there were certain problems that only a true AI could solve - and that if it could, it could also solve all others. I was also informally thinking that linguistic ability was the queen of all human skills (influenced by the Turing test itself and by the continuous failure of chatterbots). Finally, I wasn't cognisant of the possibilities of Big Data to solve these narrow problems by (clever) brute force. So I had the image of a true AI being defined by the ability to demonstrate human-like ability on linguistic problems.

Comment author: SilasBarta 17 May 2013 08:30:09PM *  4 points [-]

The game Watson was playing was non-interactive[1] -- that is, unlike with the TT, you could not change the later Jeopardy questions, based on Watson's answers, in an attempt to make it fail.

Had they done so, that would have forced an exponential blowup in the (already large) amount it would have to learn to get the same rate of correct answers.

(Not that humans would have done better in that case, of course!)

Interactivity makes a huge difference because you can focus away from its strong points and onto its weak points, thus forcing all points to be strong in order to pass.

[1] "non-adaptive" may be a more appropriate term in this context, but I say "interactive" because of the relevance of theorems about the IP complexity class (and PSPACE, which is equal).

Comment author: Stuart_Armstrong 18 May 2013 06:16:33PM 0 points [-]

I'm not saying that Watson could pass, or almost pass a Turing test. I'm saying that Watson demonstrated a combination of great quasi-linguistic skill and great general incompetence that I wasn't expecting to be possible. It proved that a computer could be "taught to the test" in at least some areas.

So I think we should keep open the possibility that a computer could be taught to the Turing test as well.

Comment author: SilasBarta 18 May 2013 07:55:41PM 0 points [-]

Well, yes, if you make the test non-adaptive, it's (exponentially) easier to pass. For example, if you limit the "conversation" to a game of chess, it's already possible. But those aren't the "full" Turing Test; they're domain-specific variants. Your criticism would only apply to the latter.

Comment author: SecondWind 26 May 2013 08:53:04AM 2 points [-]

Are AI players actually indistinguishable from humans in Chess? Could an interrogator not pick out consistent stylistic differences between equally-ranked human and AI players?