torekp comments on (Almost) every moral theory can be represented by a utility function - Less Wrong
You are viewing a comment permalink. View the original post to see all comments and the full post content.
You are viewing a comment permalink. View the original post to see all comments and the full post content.
Comments (42)
The problem with Kantianism-as-"consequentialism" is that the consequences you have to portray the agent as pursuing, are not very plausible ultimate goals, on the face of it. What makes the usual versions of consequentialism appealing is, in large part, the immediate plausibility of the claim that these goals (insert the particular theory of the good here) are what really and ultimately matter. If we specify particular types of actions in the goal (e.g., lying is to be minimized) and index to the agent, that immediate plausibility fades.