Wei_Dai comments on The Preference Utilitarian’s Time Inconsistency Problem - Less Wrong

25 Post author: Wei_Dai 15 January 2010 12:26AM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (104)

You are viewing a single comment's thread. Show more comments above.

Comment author: Wei_Dai 26 January 2010 05:34:12AM 5 points [-]

I wouldn't be so quick to discard the idea of the AI persuading us that things are pretty nice the way they are.

Suppose the AI we build (AI1) finds itself insufficiently intelligent to persuade us. It decides to build a more powerful AI (AI2) to give it advice. AI2 wakes up and modifies AI1 into being perfectly satisfied with the way things are. Then, mission accomplished, they both shut down and leave humanity unchanged.

I think what went wrong here is that this formulation of utilitarianism isn't reflectively consistent.

There are probably strong limits to the persuadability of human beings, so it wouldn't be a disaster.

If there are, then the AI would modify us physically instead.