Manfred comments on Arguments against the Orthogonality Thesis - Less Wrong

-7 Post author: JonatasMueller 10 March 2013 02:13AM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (75)

You are viewing a single comment's thread. Show more comments above.

Comment author: JonatasMueller 11 March 2013 12:24:17PM *  -1 points [-]

Indeed, a robot could be built that makes paperclips or pretty much anything. For instance, a paperclip assembling machine. That's an issue of practical implementation and not what the essay has been about, as I mention in the first paragraph and concede in the last.

The issue I argued about is that generally superintelligent agents, on their own will, without certain outside pressures from non-superintelligent agents, would understand personal identity and meta-ethics, leading them to converge to the same values and ethics. This is for two reasons: (1) they would need to take a "God's eye view" and value all perspectives besides their own, and (2) they would settle on moral realism, with the same values as good and bad feelings, in the present or future.

Comment author: Manfred 11 March 2013 10:51:49PM *  1 point [-]

Indeed, a robot could be built that makes paperclips or pretty much anything. For instance, a paperclip assembling machine. That's an issue of practical implementation and not what the essay has been about, as I mention in the first paragraph and concede in the last.

Well then, let's increase the problem to where it's meaningful, and take a look at that. What sort of cognitive and physical actions would make you think a robot is superintelligent? Discovery of new physics, modeling humans so precisely that it can predict us better than we can, making intricate plans that will work flawlessly?

What fails in the program when one tries to build a robot that takes both the paperclip-maximizing actions and superintelligent actions?

Comment author: JonatasMueller 12 March 2013 10:44:31AM *  1 point [-]

What sort of cognitive and physical actions would make you think a robot is superintelligent?

For general superintelligence, proving performance in all cognitive areas that surpasses the highest of any humans. This naturally includes philosophy, which is about the most essential type of reasoning.

What fails in the program when one tries to build a robot that takes both the paperclip-maximizing actions and superintelligent actions?

It could have a narrow superintelligence, like a calculating machine, surpassing human cognitive abilities in some areas but not in others. If it had a general superintelligence, then it would not of its own do paperclip maximization as a goal, because this would be terribly stupid, philosophically.

Comment author: Manfred 12 March 2013 03:43:24PM *  0 points [-]

it would not of its own do paperclip maximization as a goal, because this would be terribly stupid, philosophically.

My hope was to get you to support that claim in an inside-view way. Oh well.

Comment author: JonatasMueller 12 March 2013 05:20:50PM *  0 points [-]

Why it would not do paperclip (or random value) maximization as a goal is explained more at length in the article. There is more than one reason. We're considering a generally superintelligent agent, assuming above-human philosophical capacity. In terms of personal identity, there is a lack of personal identities, so it would be rational to take an objective, impersonal view, taking account of values and reasonings of relevant different beings. In terms of meta-ethics, there is moral realism and values can be reduced to the quality of conscious experience, so it would have this as its goal. If one takes moral anti-realism to be true, at least for this type of agent we are considering, a lack of real values would be understood as a lack of real goals, and could lead to the tentative goal of seeking more knowledge in order to find a real goal, or having no reason to do anything in particular (this is still susceptible to the considerations from personal identity). I argue against moral anti-realism.