Part 1 was previously posted and it seemed that people likd it, so I figured that I should post part 2 - http://waitbutwhy.com/2015/01/artificial-intelligence-revolution-2.html
Part 1 was previously posted and it seemed that people likd it, so I figured that I should post part 2 - http://waitbutwhy.com/2015/01/artificial-intelligence-revolution-2.html
There's a story about a card writing AI named Tully that really clarified the problem of FAI for me (I'd elaborate but I don't want to ruin it).
I think the fundamental misunderstanding here is that you're assuming that all intelligences are implicitly reward maximizers, even if their creators don't intend to make them reward maximizers. You, as a human, and as an intelligence based on a neural network, depend on reinforcement learning. Therefore, reward maximization is one of your many terminal values. But Bostrom proposed four other possible solutions to the value loading problem besides reinforcement learning. Here are all five in the order that they were presented in Superintelligence:
I didn't describe the last two because they're more complex, they're more tentative, I don't understand them as well, and they seem to be amalgams of the first three methods, even more so than the third method being a special case of the first.
To summarize, you thought that reward maximization was the general case because, to some extent, you're a reward maximizer. But it's actually a special case: It's not necessarily true about minds-in-general. An AI might not have a reward signal or seek to maximize one. That is to say, its terminal value(s) may not be reward maximization. I think this is what JoshuaZ was trying to get at before he started talking about wireheading.
At any rate, both kinds of AIs would result in infrastructure profusion, as JoshuaZ also seems to have implied. I don't think it matters whether it uses our atoms to make paperclips or hedonium.
But all of these things have an evaluation system in place that still comes back with a success/failure evaluation that serves as a reward/punishment system. They're different ways to use evaluative processes, but they all have pursuit of some kind positive feedback from evaluating a strategy or outcome as successful. His reinforcement learning should be called reinforcement teaching because in that one, humans are explicitly and directly in charge of the reward process whereas in the others the reward process happens more or less internally according to something that should be modifiable once the AI is sufficiently advanced.