You're looking at Less Wrong's discussion board. This includes all posts, including those that haven't been promoted to the front page yet. For more information, see About Less Wrong.

shminux comments on Approval-directed agents - Less Wrong Discussion

9 Post author: paulfchristiano 12 December 2014 10:38PM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (22)

You are viewing a single comment's thread. Show more comments above.

Comment author: shminux 12 December 2014 11:42:27PM 2 points [-]

virtue ethics, where the agent attempts to have a utility function its overseer would like?)

I don't think in virtue ethics you are obligated to maximize virtues, only satisfice them.

Comment author: gjm 13 December 2014 01:50:02PM 1 point [-]

I think maximizing versus satisficing is a question orthogonal to whether you pay attention to consequences, to hte actions that produce them, or to the character from which the actions flow. One could make a satisficing consequentialist agent, for instance. (Bostrom, IIRC, remarks that this wouldn't necessarily avoid the dangers of overzealous optimization: instead of making unboundedly many paperclips because it wants as many as possible, our agent might make unboundedly many paperclips in order to be as sure as possible that it really did make at least 10.)

Comment author: shminux 13 December 2014 10:21:50PM 0 points [-]

Boatrom's point is valid in absence of other goals. A clippy which also values a slightly non-orthogonal goal would stop making paperclips once that other goal is interfered with by the excess of paperclips.

In virtue ethics you don't maximize anything, you are free to pick any actions compatible with the virtues, so there is no utility function to speak of.