Jiro comments on The Least Convenient Possible World - Less Wrong
You are viewing a comment permalink. View the original post to see all comments and the full post content.
You are viewing a comment permalink. View the original post to see all comments and the full post content.
Comments (186)
What if we ignore the VR question? Omega tells you that killing and eating your children will make you maximally happy. Should you do it?
Omega can't tell you that doing X makes you maximally happy unless doing X actually makes you maximally happy. And a scenario where doing X actually makes you maximally happy may be a scenario where you are no longer human and don't have human preferences.
Omega could, of course, also say "you are mistaken when you conclude that being maximally happy in this scenario is not a human preference". However,