TheAncientGeek comments on Fake Utility Functions - Less Wrong

22 Post author: Eliezer_Yudkowsky 06 December 2007 04:55PM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (54)

Sort By: Old

You are viewing a single comment's thread. Show more comments above.

Comment author: DefectiveAlgorithm 19 June 2014 02:02:45PM *  0 points [-]

a terminal goal of interpreting instructions correctly

There is a huge amount of complexity hidden beneath this simple description.

Comment author: TheAncientGeek 19 June 2014 05:01:02PM *  -1 points [-]

I'll say it again: absolute complexity is not relative complexity.

Everything in AGI us very complex in absolute teams.

In relative terms, language is less complex than language+morality

Comment author: VAuroch 23 August 2014 09:33:26PM -1 points [-]

That would matter if you didn't need language+morality to interpret language in this case. To interpret instructions correctly, you have to understand what they mean, and that requires a full understanding of the motivations underlying the request.

You don't just need language, you need language+thought, which is even more complex than language+morality.

Comment author: TheAncientGeek 24 August 2014 04:29:20PM 1 point [-]

I am using "having language" to mean "having language plus thought", ie to have linguistic understanding, ie to have the ability to pass a Turning Test. Language without thought is just parotting.

To follow instructions relating morality correctly, an entity must be able to understand them correctly at the semantic levlel. An entity need not agree with them, or hold to them itself, as we can see from the ability of people to play along with social rules they don't personally agree with.

Comment author: VAuroch 24 August 2014 10:23:01PM *  -1 points [-]

No, that's not right. language + thought is to understand language and be able to fully model the mindstate of the person who was speaking to you. If you don't have this, and just have language, 'get grandma out of the burning house ' gets you the lethal ejector seat method. If you want do-what-I-mean rather than do-what-I-say, you need full thought modeling. Which is obviously harder than language + morality, which requires only being able to parse language correctly and understand a certain category of thought.

Or to phrase it a different way: language on its own gets you nothing productive, just a system that can correctly parse statements. To understand what they mean, rather than what they say, you need something much broader, and language+morality is smaller than that broad thing.

Comment author: TheAncientGeek 26 August 2014 05:34:39PM *  -1 points [-]

Fully understanding the semantics of morality may be simpler than fully understanding the semantics of everything, but it doesn't get you AI safety, because an AI can understand something without being motivated to act on it.

When I wrote "language", I meant words + understanding ....understanding in general, therefore including understanding of ethics..and when I wrote "morality" I meant a kind motivation.