lukeprog comments on (Almost) every moral theory can be represented by a utility function - Less Wrong

5 Post author: lukeprog 30 April 2012 03:31AM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (42)

You are viewing a single comment's thread. Show more comments above.

Comment author: lukeprog 01 May 2012 05:47:08PM 0 points [-]

Who is weakening the conditions for a theory being "consequentialism"? The thing described by Peterson seems perfectly in line with consequentialism. And his point about asymmetry among moral theories remains.

Comment author: bryjnar 02 May 2012 11:17:56AM 1 point [-]

Well, there are a lot of things that get called "consequentialism" (take a look at the SEP article for a similar point). I personally find that "consequentialism" connotes to "agent-neutral" in my head, but that may just be me. I feel like requiring neutrality is a more interesting position precisely because bare consequentialism is so weak: it's not really surprising that almost everything is a form of it.

There's also the possibility of accidental equivocation, since people use "consequentialism" to stand for so many things. I actually think the stronger interpretations are pretty common (again, the SEP article has a little discussion on this), and so there is some danger of people thinking that this shows a stronger result than it actually does.

Comment author: lukeprog 02 May 2012 07:16:12PM 0 points [-]

Nah, people argue all the time about agent neutrality. Agent-neutral consequentialism is simply one form of consequentialism, albeit a popular one.

Comment author: torekp 01 May 2012 11:44:36PM *  1 point [-]

The problem with Kantianism-as-"consequentialism" is that the consequences you have to portray the agent as pursuing, are not very plausible ultimate goals, on the face of it. What makes the usual versions of consequentialism appealing is, in large part, the immediate plausibility of the claim that these goals (insert the particular theory of the good here) are what really and ultimately matter. If we specify particular types of actions in the goal (e.g., lying is to be minimized) and index to the agent, that immediate plausibility fades.