Vladimir_Nesov comments on David Chalmers' "The Singularity: A Philosophical Analysis" - Less Wrong

33 Post author: lukeprog 29 January 2011 02:52AM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (202)

You are viewing a single comment's thread. Show more comments above.

Comment author: Vladimir_Nesov 30 January 2011 03:53:30AM 0 points [-]

To get to that point we have to start from the right meaning to begin with, and care about preserving it accurately, and Jacob doesn't agree those steps are important or particularly hard.

Comment author: jacob_cannell 30 January 2011 04:21:55AM -1 points [-]

Not quite.

As for the start with the right meaning part, I think it is extremely hard to 'solve' morality in the way typically meant here with CEV or what not.

I don't think that we need (or will) wait to solve that problem before we build AGI, any more or less than we need to solve it for having children and creating a new generation of humans.

If we can build AGI somewhat better than us according to our current moral criteria, they can build an even better successive generation, and so on - a benevolence explosion.

As for the second part about preserving it accurately, I think that ethics/morality is complex enough that it can only be succinctly expressed in symbolic associative human languages. An AGI could learn how to model (and value) the preferences of others in much the same way humans do.

Comment author: wedrifid 30 January 2011 04:27:22AM 3 points [-]

I don't think that we need (or will) wait to solve that problem before we build AGI, any more or less than we need to solve it for having children and creating a new generation of humans.

If we can build AGI somewhat better than us according to our current moral criteria, they can build an even better successive generation, and so on - a benevolence explosion.

Someone help me out. What is the right post to link to that goes into the details of why I want to scream "No! No! No! We're all going to die!" in response to this?

Comment author: Vladimir_Nesov 30 January 2011 09:19:17AM 0 points [-]

Coming of Age sequence examined realization of this error from Eliezer's standpoint, and has further links.

Comment author: jacob_cannell 30 January 2011 10:16:54AM 0 points [-]

In which post? I'm not finding discussion about the supposed danger of improved humanish AGI.

Comment author: Vladimir_Nesov 30 January 2011 10:22:32AM *  -1 points [-]

That Tiny Note of Discord, say. (Not on "humanish" AGI, but eventually exploding AGI.)

Comment author: jacob_cannell 30 January 2011 07:07:34PM *  0 points [-]

I don't see much of a relation at all to what i've been discussing in that first post.

[http://lesswrong.com/lw/lq/fake_utility_functions/] is a little closer, but still doesn't deal with human-ish AGI.

Comment author: nshepperd 30 January 2011 06:05:05AM *  2 points [-]

Why would an AI which optimises for one thing create another AI that optimises for something else? Not every change is an improvement, but every improvement is necessarily a change. Building an AI with a different utility function is not going to satisfy the first AI's utility function! So whatever AI the first one builds is necessarily going to either have the same utility function (in which case the first AI is working correctly), or have a different one (which is a sign of malfunction, and given the complexity of morality, probably a fatal one).

It's not possible to create an AGI that is "somewhat better than us" in the sense that it has a better utility function. To the extent that we have a utility function at all, it would refer to the abstract computation called "morality", which "better" is defined by. The most moral AI we could create is therefore one with precisely that utility function. The problem is that we don't exactly know what our utility function is (hence CEV).

There is a sense in which a Friendly AGI could be said to be "better than us", in that a well-designed one would not suffer from akrasia and whatever other biases prevent us from actually realizing our utility function.

Comment author: Stuart_Armstrong 30 January 2011 06:17:57PM 2 points [-]

AI's without utility functions, but some other motivational structure, will tend to self-improve to a utility function AI. Utility-function AI's seem more stable under self-improvement, but there are many reasons it might want to change its utility (eg speed of access, multi-agent situations).

Comment author: Oligopsony 30 January 2011 06:53:26PM 0 points [-]

Could you clarify what you mean by an "other motivational structure?" Something with preference non-transitivity?

Comment author: Stuart_Armstrong 30 January 2011 07:47:18PM 1 point [-]
Comment author: Perplexed 30 January 2011 05:04:21PM 2 points [-]

Why would an AI which optimises for one thing create another AI that optimises for something else?

It wouldn't if it initially considered itself to be the only agent in the universe. But if it recognizes the existence of other agents and the impact of other agents' decisions on its own utility, then there are many possibilities:

  • The new AI could be created as a joint venture of two existing agents.
  • The new AI could be built because the builder was compensated for doing so.
  • The new AI could be built because the builder was threatened into doing so.

Building an AI with a different utility function is not going to satisfy the first AI's utility function!

This may seem intuitively obvious, but it is actually often false in a multi-agent environment.