Stuart_Armstrong comments on Siren worlds and the perils of over-optimised search - Less Wrong

27 Post author: Stuart_Armstrong 07 April 2014 11:00AM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (411)

You are viewing a single comment's thread. Show more comments above.

Comment author: Stuart_Armstrong 07 May 2014 04:52:04PM 0 points [-]

No, it's preferences the problem, not understanding. Why would an AI sitting at the end of a series of self improvements choose to interpret ambiguous coding in the way we prefer?

I can say that if instrumental rationality is dangerous, don't bulld it that way.

How do you propose to build an AI without instrumental rationality or preventing that from developing? And how do you propose to convince AI designers to go down that route?

Comment author: TheAncientGeek 07 May 2014 05:31:43PM *  0 points [-]

If it has epistemic rationality as a goal, it will default to getting things right rather than wrong.

Not only nstrumental rationality = epistemic rationality.

Comment author: Stuart_Armstrong 12 May 2014 11:05:10AM *  0 points [-]

If it has epistemic rationality as a goal, it will default to getting things right rather than wrong.

If it has epistemic rationality as a goal, it will default to acquiring true beliefs about the world. Explain how this will make it "nice".

Comment author: TheAncientGeek 12 May 2014 12:16:41PM *  0 points [-]

See above. The question was originally about interpreting directives. You have switched to inferring morality apriori.