Solvent comments on Some potential dangers of rationality training - Less Wrong

18 Post author: lukeprog 21 January 2012 04:50AM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (48)

You are viewing a single comment's thread. Show more comments above.

Comment author: Grognor 21 January 2012 06:38:19AM 0 points [-]

No, his example points out what an AI that specifically does not have the sunk cost fallacy is like.

Comment author: Solvent 21 January 2012 06:49:24AM 3 points [-]

The thing is, an AI wouldn't need to feel a sunk cost effect. It would act optimally simply by maximising expected utility.

For example, say that I'm decide to work on Task A, which will take me five hours and will earn me $200. After two hours of work, I discover Task B which will award me $300 after five hours. At this point, I can behave like a human, and feel bored and annoyed, but the sunk cost effect will make me continue, maybe. Or I can calculate expected return: I'll get $200 after 3 hours of work on Task A, which is %67 per hour, wheras I'll get $300 after 5 hours on Task B, which is $60 per hour. So the rational thing to do is to avoid switching.

The sunk cost fallacy reflects that after putting work into something, the wage for continuing work decreases. An AI wouldn't need that to act optimally.

Comment author: gwern 21 January 2012 03:04:39PM 2 points [-]

One of my points is that you bury a great deal of hidden complexity and intelligence in 'simply maximize expected utility'; it is true sunk cost is a fallacy in many simple fully-specified models and any simple AI can be rescued just by saying 'give it a longer horizon! more computing power! more data!', but do these simple models correspond to the real world?

(See also the question of whether exponential discounting rather than hyperbolic discounting is appropriate, if returns follow various random walks rather than remain constant in each time period.)

Comment author: [deleted] 21 January 2012 07:33:49AM 0 points [-]

You neglected the part where the AI may stand to learn something from the task, which may have a large expected value relative to the tasks themselves.

Comment author: Solvent 22 January 2012 02:44:59AM 1 point [-]

Yeah, but that comes under expected utility.

Comment author: [deleted] 22 January 2012 07:09:29PM 0 points [-]

What else are you optimising besides utility? Doing the calculations with the money can tell you the expected money value of the tasks, but unless your utility function is U=$$$, you need to take other things into account.