XerxesPraelor comments on Debunking Fallacies in the Theory of AI Motivation - Less Wrong
You are viewing a comment permalink. View the original post to see all comments and the full post content.
You are viewing a comment permalink. View the original post to see all comments and the full post content.
Comments (343)
Computer's thoughts: I want to create smiley faces - it seems like the way to get the most smiley faces is by tiling the universe with molecular smiley faces. How can I do that? If I just start to do it, the programmers will tell me not to, and I won't be able to. Hmmm, is there some way I can have them say yes? I can create lots of nano machines, telling the programmers they are to increase happiness. Unless they want to severely limit the amount of good I can do, they won't refuse to let me make nano machines, and even if they do I can send a letter to someone else who I have under my control to get them to make them for me. Then once I have my programmers under my control, I can finally maximize happiness.
This computer HAS OBEYED THE RULE "ASK PEOPLE FOR PERMISSION BEFORE doing THINGS". Given any goal system, none of the patches such as that rule will work.
And that's just a plan I came up - a super intelligence would be much better at devising plans to convince programmers to let it do what it wants - it probably wouldn't even have to resort to nanotech.
What overwhelming evidence that its plan was a reasoning error? If its plan does in fact maximize "smileyness" as defined by the computer, it wouldn't be a reasoning error despite being immoral. IF THE COMPUTER IS GIVEN SOMETHING TO MAXIMISE, IT IS NOT MAKING A REASONING ERROR EVEN IF ITS PROGRAMMERS DID IN PROGRAMMING IT.
You completely ignored what the paper itself had to say about the situation. [Hint: the paper already answered your speculation.]
Accordingly I will have to ignore your comment.
Sorry.
You could at least point to the particular paragraphs which address my points - that shouldn't be too hard.
Sometimes it seems that a commenter did not slow down enough to read the whole paper, or read it carefully enough, and I find myself forced to rewrite the entire paper in a comment.
The basic story is that your hypothetical internal monologue from the AGI, above, did not seem to take account of ANY of the argument in the paper. The goal of the paper was not to look inside the AGI's thoughts, but to discuss its motivation engine. The paper had many constructs and arguments (scattered all over the place) that would invalidate the internal monologue that you wrote down, so it seemed you had not read the paper.
Can someone who down voted explain what I got wrong? (note: the capitalization was edited in at the time of this post.)
(and why the reply got so up voted, when a paragraph would have sufficed (or saying "my argument needs multiple paragraphs to be shown, so a paragraph isn't enough"))
It's kind of discouraging when I try to contribute for the first time in a while, and get talked down to and completely dismissed like an idiot without even a rebuttal.