You're looking at Less Wrong's discussion board. This includes all posts, including those that haven't been promoted to the front page yet. For more information, see About Less Wrong.

paulfchristiano comments on Shane Legg's Thesis: Machine Superintelligence, Opinions? - Less Wrong Discussion

9 Post author: Zetetic 08 May 2011 08:04PM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (45)

You are viewing a single comment's thread. Show more comments above.

Comment author: paulfchristiano 09 May 2011 04:56:57AM *  3 points [-]

In my view, all results should be expressible in the form: because the real world data exhibits empirical structure X, algorithm Y succeeds in describing/predicting it.

Even if I know the exact probability distribution over images, there is an algorithmic problem (namely, how to do the inference), so your view is definitely at least a little too extreme.

In fact, this algorithmic difficulty is an issue that many researchers are currently grappling with, so in practice you really shouldn't expect all results to be making a novel statement about how the world works. Applying this standard to current research would stall progress in the directions I (and I think most serious AI researchers) currently believe are most important to actually reaching AI, especially human-comprehensible AI which might possibly be friendly.

Maybe we are wrong, but the argument you gave, and your implications about how NFL should be applied, are not really relevant to that question.

Comment author: Daniel_Burfoot 09 May 2011 04:52:08PM 0 points [-]

Even if I know the exact probability distribution over images, there is an algorithmic problem (namely, how to do the inference), so your view is definitely at least a little too extreme.

I don't dispute that the algorithmic problem is interesting and important. I only claim that the empirical question is equally important.

Applying this standard to current research would stall progress in the directions I (and I think most serious AI researchers) currently believe are most important to actually reaching AI

What you're really saying is that you think a certain direction of research will be fruitful. That's fine. I disagree, but I doubt we can resolve the debate. Let's compare notes again in 2031.