novalis comments on Safe questions to ask an Oracle? - Less Wrong

2 Post author: Stuart_Armstrong 27 January 2012 06:33PM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (41)

You are viewing a single comment's thread. Show more comments above.

Comment author: novalis 27 January 2012 10:45:16PM -1 points [-]

Do you think this is a loophole allowing arbitrary actions? Or do you think that an AI would simply say, "I don't know what it means for energy to be consumed, so I'm not going to do anything."

I don't know much about physics, but do you think that some sort of measure of entropy might work better?

Comment author: saturn 28 January 2012 12:19:04AM *  1 point [-]

As far as I know, every simple rule either leaves trivial loopholes, or puts the AI on the hook for a large portion of all the energy (or entropy) in its future light cone, a huge amount which wouldn't be meaningfully related to how much harm it can do.

If there is a way around this problem, I don't claim to be knowledgeable or clever enough to find it, but this idea has been brought up before on LW and no one has come up with anything so far.

Comment author: novalis 28 January 2012 12:53:56AM 0 points [-]

Link to previous discussions?

It seems that many questions could be answered using only computing power (or computing power and network access to more-or-less static resources), and this doesn't seem like a difficult limitation to put into place. We're already assuming a system that understands English at an extremely high level. I'm convinced that ethics is hard for machines because it's hard for humans too. But I don't see why, given an AI worthy of the name, following instructions is hard, especially given the additional instructions, "be conservative and don't break the law".

Comment author: orthonormal 28 January 2012 01:03:39AM 0 points [-]

Link to previous discussions?

Dreams of Friendliness