Emile comments on Open Thread, July 1-15, 2012 - Less Wrong
You are viewing a comment permalink. View the original post to see all comments and the full post content.
You are viewing a comment permalink. View the original post to see all comments and the full post content.
Comments (150)
If that first agent (that answers no, then self-modifies to answer yes) had been in the situation where the coin had fell heads, then it would not have got the million dollars; whereas an agent that can "retroactively precommit" to answer yes would have got the million dollars. So having a "retroactively precommit" algorithm seems like a better choice than having a "answer what gets the biggest reward, and then self-modify for future cases" algorithm.
But we know that didn't happen. Why do we care about utility we know we can't obtain?
For what goal is this a better choice? Utility generation?