pnrjulius comments on SotW: Check Consequentialism - Less Wrong

38 Post author: Eliezer_Yudkowsky 29 March 2012 01:35AM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (311)

You are viewing a single comment's thread. Show more comments above.

Comment author: pnrjulius 04 April 2012 08:44:06PM 0 points [-]

In game theory, there are a number of situations where it is rational to handicap your own rationality: Reduce your number of choices, take away information, etc.

Now, in game theory you're competing against someone else, whereas in this case you're only competing against (time-indexed versions of?) yourself; but it could be that the same rules apply. Maybe it really is rational to pay to not know something.

Or maybe it's rational for a bounded agent to pay to be counter-biased: Knowing that I have this bias toward sunk costs, make me ignorant of all sunk costs.

Comment author: Eliezer_Yudkowsky 05 April 2012 08:43:56PM 3 points [-]

In game theory, there are a number of situations where it is rational to handicap your own rationality: Reduce your number of choices, take away information, etc.

TDT is intended to eliminate this. A TDT-agent - one that's correctly modeled by the environment, not that some other agent thinks is a CDT-agent - is supposed to never benefit from having any option taken away from it, and will never pay to avoid learning a piece of information.

Comment author: jeremysalwen 05 April 2012 08:49:11PM 2 points [-]

Er, this is assuming that the information revealed is not intentionally misleading, correct? Because certainly you could give a TDT agent an extra option which would be rational to take on the basis of the information available to the agent, but which would still be rigged to be worse than all other options.

Or in other words, the TDT agent can never be aware of such a situation.

Comment author: Eliezer_Yudkowsky 05 April 2012 09:22:28PM 1 point [-]

Amendment accepted.

Comment author: handoflixue 04 April 2012 11:09:23PM 0 points [-]

Agreed. I think one could assert "Given a perfect decision theory AND a perfect implementation, additional information is never a negative", but it's silly to live as though that were true. If you know your decision theory doesn't handle X information correctly (say, sunken costs) then it's in your best interests to either eliminate the information, or fix the decision theory.

Of course, eliminating information seems to be by far the easier option...

Comment author: TheOtherDave 04 April 2012 11:25:36PM 1 point [-]

If I know the class of errors my decision theory tends to make given the kinds of Xes I most commonly run into, I can also adopt a third option... for want of a better term, I can patch my decision theory. E.g., "Well, I want to finish this project, but I suspect that part of that desire stems from an invalid weighting of sunk costs, so I won't take that desire at face value... I'll apply some kind of rough-and-ready discounting factor to it." This is clearly not as good as actually fixing my decision theory, but isn't as hard either, and is sometimes more practical than eliminating the information.

Comment author: handoflixue 05 April 2012 06:45:39PM 0 points [-]

Very true. However, "avoid X information, since it biases me" is actually an example of such a patch. Especially if the information doesn't otherwise have any useful value. How often does knowledge of sunk costs actually move you towards ideal action, rather than biasing you away from it?

Comment author: TheOtherDave 05 April 2012 07:14:40PM 0 points [-]

Sure, avoiding information is an example of patching a decision theory, agreed.

So I guess what I'm saying is that "either eliminate the information, or fix the decision theory" is a misleading way to phrase the choice. My real choice is between fixing it and patching it, where eliminating the information is one of several ways to patch it, and not always the best.

Making choices about future investments in ignorance of the existing data I have about previous investments and their ROI is probably less ideal than taking those data into consideration and applying some other patch to compensate for sunk-costing.

Comment author: handoflixue 05 April 2012 10:56:52PM 0 points [-]

I like the idea of phrasing it as "patching vs long-term fixes" :)