Today's post, Morality as Fixed Computation was originally published on 08 August 2008. A summary (taken from the LW wiki):
A clarification about Yudkowsky's metaethics.
Discuss the post here (rather than in the comments to the original post).
This post is part of the Rerunning the Sequences series, where we'll be going through Eliezer Yudkowsky's old posts in order so that people who are interested can (re-)read and discuss them. The previous post was Contaminated by Optimism, and you can use the sequence_reruns tag or rss feed to follow the rest of the series.
Sequence reruns are a community-driven effort. You can participate by re-reading the sequence post, discussing it here, posting the next day's sequence reruns post, or summarizing forthcoming articles on the wiki. Go here for more details, or to have meta discussions about the Rerunning the Sequences series.
This comment by Toby Ord is nice:
The question gets especially interesting if we imagine that Mary's sensitivity to moral arguments is due to a bug in her brain, and otherwise she's completely on board with her original selfish desires.
Well, the trouble is that we're not just not utility-maximizers, we're not even stable under small perturbations :)