You're looking at Less Wrong's discussion board. This includes all posts, including those that haven't been promoted to the front page yet. For more information, see About Less Wrong.

buybuydandavis comments on Examples of AI's behaving badly - Less Wrong Discussion

25 Post author: Stuart_Armstrong 16 July 2015 10:01AM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (35)

You are viewing a single comment's thread. Show more comments above.

Comment author: buybuydandavis 16 July 2015 11:38:27PM 2 points [-]

I see this failure in analysis all the time.

When people want to change the behavior of others, they find some policy and incentive that would encourage the change they desire, but never stop to ask how else people might react to that change in incentives.

Anyone ever come across any catchy name or formulation for this particular failure mode?

Comment author: ThisSpaceAvailable 17 July 2015 05:55:54AM 7 points [-]

Perverse incentives.

Comment author: gjm 17 July 2015 12:31:09PM 3 points [-]

Cobra effect (see the Wikipedia page I linked before). Law of unintended consequences. (Perhaps the former is a little too narrow and the latter a little too broad.)

Comment author: robertzk 02 August 2015 02:27:41AM 1 point [-]

Isn't this an example of a reflection problem? We induce this change in a system, in this case an evaluation metric, and now we must predict not only the next iteration but the stable equilibria of this system.

Comment author: bbleeker 17 July 2015 03:55:56PM *  0 points [-]

Goodhart's Law

Oops, double post; V_V already said that.

Comment author: [deleted] 17 July 2015 02:17:11AM 0 points [-]

I believe this is called a "red queen race"

Comment author: Toggle 17 July 2015 01:59:27PM 0 points [-]

This is not correct, at least in common usage.

A Red Queen's Race is an evolutionary competition in which absolute position does not change. The classic example is the arms race between foxes and rabbits that results in both becoming faster in absolute terms, but the rate of predation stays fixed. (The origin is Lewis Carrol: "It takes all the running you can do, just to stay in the same place.")

Comment author: Lumifer 17 July 2015 02:53:33PM 2 points [-]

A Red Queen's Race is an evolutionary competition in which absolute position does not change.

You mean relative, not absolute.

I've also seen a more general interpretation: the Red Queen situation is where staying still (doing nothing) makes you worse off as time passes; you need to run forward just to stay in the same place.

Comment author: Toggle 17 July 2015 04:59:41PM 1 point [-]

You mean relative, not absolute.

Yes, yes I did. Thanks for the correction.

Comment author: [deleted] 17 July 2015 11:13:10PM 0 points [-]

I think this is analogous to what's happening here - you create better incentives, they create better ways to get around those incentives, nothing changes. I didn't know that this wasn't the common usage, as I got it from this Overcoming Bias post:

http://www.overcomingbias.com/2014/06/bias-is-a-red-queen-game.html