Wei_Dai comments on Will the ems save us from the robots? - Less Wrong

8 Post author: Stuart_Armstrong 24 November 2011 07:23PM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (43)

You are viewing a single comment's thread. Show more comments above.

Comment author: Wei_Dai 12 March 2012 08:17:51PM *  2 points [-]

Using the Socratic method to reconstruct standard WBE analysis, which I think we are both already familiar with, is a red herring.

Is the standard WBE analysis written up anywhere? By that phrase do you mean to include the "number of person-months" of work by FHI/SIAI that you mentioned earlier? I really am uncertain how far FHI/SIAI has pushed the analysis in these areas, and my questions were meant to be my attempt to figure that out. But it does seem like most of our disagreement is over decision theory rather than WBE, so let's move the focus there.

Most have seemed to think that decision theory is a very small piece of the AGI picture.

I also think that's most likely the case, but there's a significant chance that it isn't. I have not heard a strong argument why decision theory must be a very small piece of the AGI picture (and I did bring up this question on the decision theory mailing list), and in my state of ignorance it doesn't seem crazy to think that maybe with the right decision theory and just a few other key pieces of technology, AGI would be possible.

As for thinking there is ~0 chance of an FAI research program leading to safe AI, my reasoning is that with FAI we're dealing with seemingly impossible problems like ethics and consciousness, as well as numerous other philosophical problems that aren't quite thousand-years old, but still look quite hard. What are the chances all these problems get solved in a few decades, barring IA and WBE? If we do solve them, we still have to integrate the solutions into an AGI design, verify its correctness, avoid Friendliness-impacting implementation bugs, and do all of that before other AGI projects take off.

It's the social consequences that I'm most unsure about. It seems like if SIAI can keep "ownership" over the decision theory ideas and use it to preach AI risk, then that would be beneficial, but it could also be the case that the ideas take on a life of their own and we just end up having more people go into decision theory because they see it as a fruitful place to get interesting technical results.

Comment author: timtyler 12 March 2012 11:01:10PM *  -2 points [-]

Most have seemed to think that decision theory is a very small piece of the AGI picture.

I also think that's most likely the case, but there's a significant chance that it isn't. I have not heard a strong argument why decision theory must be a very small piece of the AGI picture (and I did bring up this question on the decision theory mailing list), and in my state of ignorance it doesn't seem crazy to think that maybe with the right decision theory and just a few other key pieces of technology, AGI would be possible.

On one hand, machine intelligence is all about making decisions in the face of uncertainty - so from this perspective, decision theory is central.

On the other hand, the basics of decision theory do not look that complicated - you just maximise expected utility. The problems seem to be mostly down to exactly how to do that efficiently.

The idea that safe machine intelligence will be assisted by modifications to decision theory to deal with "esoteric" corner cases seems to be mostly down to Eliezer Yudkowsky. I think it is a curious idea - but I am very happy that it isn't an idea that I am faced with promoting.

Comment author: Wei_Dai 12 March 2012 11:12:17PM 3 points [-]

On the other hand, the basics of decision theory do not look that complicated - you just maximise expected utility. The problems seem to be mostly down to exactly how to do that efficiently.

Isn't AIXI a counter-example to that? We could give it unlimited computing power, and it would still screw up badly, in large part due to a broken decision theory, right?

Comment author: timtyler 12 March 2012 11:35:44PM *  0 points [-]

Kinda, yes. Any problem is a decision theory problem - in a sense. However, we can get a long way without the wirehead problem, utility counterfeiting, and machines mining their own brains causing problems.

From the perspective of ordinary development these don't look like urgent issues - we can work on them once we have smarter minds. We need not fear not solving them too much - since if we can't solve these problems our machines won't work and nobody will buy them. It would take security considerations to prioritise these problems at this stage.