cousin_it comments on The genie knows, but doesn't care - Less Wrong

54 Post author: RobbBB 06 September 2013 06:42AM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (515)

You are viewing a single comment's thread. Show more comments above.

Comment author: cousin_it 10 January 2014 12:47:30PM *  3 points [-]

I don't understand your comment, and I no longer understand your grandparent comment either. Are you using a meaning of "morality" that is distinct from "preferences"? If yes, can you describe your assumptions in more detail? It's not just for my benefit, but for many others on LW who use "morality" and "preferences" interchangeably.

Comment author: ArisKatsaris 10 January 2014 12:56:49PM 1 point [-]

but for many others on LW who use "morality" and "preferences" interchangeably.

Do that many people really use them interchangeably? Would these people understand the questions "Do you prefer chocolate or vanilla ice-cream?" as completely identical in meaning to "Do you consider chocolate or vanilla as the morally superior flavor for ice-cream?"

Comment author: cousin_it 10 January 2014 01:17:59PM *  6 points [-]

I don't care about colloquial usage, sorry. Eliezer has a convincing explanation of why wishes are intertwined with morality ("there is no safe wish smaller than an entire human morality"). IMO the only sane reaction to that argument is to unify the concepts of "wishes" and "morality" into a single concept, which you could call "preference" or "morality" or "utility function", and just switch to using it exclusively, at least for AI purposes. I've made that switch so long ago that I've forgotten how to think otherwise.

Comment author: [deleted] 10 January 2014 03:45:39PM 1 point [-]

I recommend you re-learn how to think otherwise so you can fool humans into thinking you're one of them ;-).

Comment author: TheAncientGeek 13 January 2014 02:50:13PM *  0 points [-]

I don't care about colloquial usage, sorry. e You should car, because no-one can make valid arguments based on arbitrary definitions. I can't prove angels exist, by redefining "angel" to mean what "seagull" means. How can you tell when a redefinition is arbitrary (since there are legitmate redefinitions)? Too much departure from colloquial usage.

Eliezer has a convincing explanation of why wishes are intertwined with morality ("there is no safe wish smaller than an entire human morality").

"Intertwined with" does not mean "the same as".

I am not convinced by the explanation. It also applies ot non-moral prefrences. If I have a lower priority non moral prefence to eat tasty food, and a higher priority preference to stay slim, I need to consider my higher priority preferece when wishing for yummy ice cream.

To be sure, an agent capable of acting morally will have morality among their higher priority preferences -- it has to be among the higher order preferences, becuase it has to override other preferences for the agent to act morally. Therefore, when they scan their higher prioriuty prefences, they will happen to encounter their moral preferences. But that does not mean any preference is necessarily a moral preference. And their moral prefences override other preferences which are therefore non-moral, or at least less moral.

Therefore morality si a subset of prefences, as common sense maintained all along.

I've made that switch so long ago that I've forgotten how to think otherwise.

IMO, it is better to keep ones options open.

Comment author: ArisKatsaris 10 January 2014 02:24:40PM *  0 points [-]

I don't experience the emotions of moral outrage and moral approval whenever any of my preferences are hindered/satisfied -- so it seems evident that my moral circuitry isn't identical to my preference circuitry. It may overlap in parts, it may have fuzzy boundaries, but it's not identical.

My own view is that morality is the brain's attempt to extrapolate preferences about behaviours as they would be if you had no personal stakes/preferences about a situation.

So people don't get morally outraged at other people eating chocolate icecreams, even when they personally don't like chocolate icecreams, because they can understand that's a strictly personal preference. If they believe it to be more than personal preference and make it into e.g. "divine commandment" or "natural law", then moral outrage can occur.

That morality is a subjective attempt at objectivity explains many of the confusions people have about it.

Comment author: [deleted] 10 January 2014 07:11:53PM 0 points [-]

The ice cream example is bad because the consequences are purely internal to the person consuming the ice cream. What if the chocolate ice cream was made with slave labour? Many people would then object to you buying it on moral grounds.

Eliezer has produced an argument I find convincing that morality is the back propagation of preference to the options of an intermediate choice. That is to say, it is "bad" to eat chocolate ice cream because it economically supports slavers, and I prefer a world without slavery. But if I didn't know about the slave-labour ice cream factory, my preference would be that all-things-being-equal you get to make your own choices about what you eat, and therefore I prefer that you choose (and receive) the one you want, which is your determination to make, not mine.

Do you agree with EY's essay on the nature of right-ness which I linked to?

Comment author: cousin_it 10 January 2014 02:40:50PM *  0 points [-]

my moral circuitry isn't identical to my preference circuitry

That doesn't seem to be required for Eliezer's argument...

I guess the relevant question is, do you think FAI will need to treat morality differently from other preferences?

Comment author: ArisKatsaris 10 January 2014 03:29:53PM 0 points [-]

Do you think FAI will need to treat morality differently from other preferences?

I would prefer a AI that followed my extrapolated preferences, than a AI that followed my morality. But a AI that followed my morality would be morally superior to an AI that followed my extrapolated preferences.

If you don't understand the distinction I'm making above, consider a case of the AI having to decide whether to save my own child vs saving a thousand random other children. I'd prefer the former, but I believe the latter would be the morally superior choice.

Is that idea really so hard to understand? Would you dismiss the distinction I'm making as merely colloquial language?

Comment author: [deleted] 10 January 2014 08:11:54PM *  2 points [-]

If you don't understand the distinction I'm making above, consider a case of the AI having to decide whether to save my own child vs saving a thousand random other children. I'd prefer the former, but I believe the latter would be the morally superior choice.

Wow there is so much wrapped up in this little consideration. The heart of the issue is that we (by which I mean you, but I share your delimma) have truly conflicting preferences.

Honestly I think you should not be afraid to say that saving your own child is the moral thing to do. And you don't have to give excuses either - it's not that “if everyone saved their own child, then everyone's child will be looked after” or anything like that. No, the desire to save your own child is firmly rooted in our basic drives and preferences, enough so that we can go quite far in calling it a basic foundational moral axiom. It's not actually axiomatic, but we can safely treat it as such.

At the same time we have a basic preference to seek social acceptance and find commonality with the people we let into our lives. This drives us to want outcomes that are universally or at least most-widely acceptable, and seek moral frameworks like utilitarianism which lead to these outcomes. Usually this drive is secondary to self-serving preferences for most people, and that is OK.

For some reason you've called making decisions in favor of self-serving drives "preferences" and decisions in favor of social drives "morality." But the underlying mechanism is the same.

"But wait, if I choose self-serving drives over social conformity, doesn't that lead to me to make the decision to save one life in exclusion to 1000 others?" Yes, yes it does. This massive sub-thread started with me objecting to the idea that some "friendly" AI somewhere could derive morality experimentally from my preferences or the collective preferences of humankind, make it consistent, apply the result universally, and that I'd be OK with that outcome. But that cannot work because there is not, and cannot be a universal morality that satisfies everyone - every one of those thousand other children have parents that want their kid to survive and would see your child dead if need be.

Comment author: ArisKatsaris 10 January 2014 08:29:16PM 0 points [-]

Honestly I think you should not be afraid to say that saving your own child is the moral thing to do

What do you mean by "should not"?

and that is OK.

What do you mean by "OK"?

For some reason you've called making decisions in favor of self-serving drives "preferences" and decisions in favor of social drives "morality." But the underlying mechanism is the same.

Show me the neurological studies that prove it.

But that cannot work because there is not, and cannot be a universal morality that satisfies everyone - every one of those thousand other children have parents that want their kid to survive and would see your child dead if need be.

Yes, and yet if none of the children were mine, and if I wasn't involved in the situation at all, I would say "save the 1000 children rather than the 1". And if someone else, also not personally involved, could make the choice and chose to flip a coin instead in order to decide, I'd be morally outraged at them.

You can now give me a bunch of reasons of why this is just preference, while at the same time EVERYTHING about it (how I arrive to my judgment, how I feel about the judgment of others) makes it a whole distinct category of its own. I'm fine with abolishing useless categories when there's no meaningful distinction, but all you people should stop trying to abolish categories where there pretty damn obviously IS one.

Comment author: blacktrance 10 January 2014 08:42:04PM 1 point [-]

What do you mean by "should not"?

I suspect that he means something like 'Even though utilitarianism (on LW) and altruism (in general) are considered to be what morality is, you should not let that discourage you from asserting that selfishly saving your own child is the right thing to do". (Feel free to correct me if I'm wrong.)

Comment author: [deleted] 10 January 2014 08:45:05PM 1 point [-]

Yes that is correct.

Comment author: ArisKatsaris 10 January 2014 08:59:17PM 0 points [-]

So you explained "should not" by using a sentence that also has "should not" in it.

Comment author: [deleted] 10 January 2014 08:35:05PM 0 points [-]

I'm fine with abolishing useless categories when there's no meaningful distinction, but all you people should stop trying to abolish categories where there pretty damn obviously IS one.

I've explained to you twice now how the two underlying mechanisms are unified, and pointed to Eliezer's quite good explanation on the matter. I don't see the need to go through that again.

Comment author: cousin_it 10 January 2014 03:59:19PM 0 points [-]

I would prefer a AI that followed my extrapolated preferences, than a AI that followed my morality. But a AI that followed my morality would be morally superior to an AI that followed my extrapolated preferences.

If you were offered a bunch of AIs with equivalent power, but following different mixtures of your moral and non-moral preferences, which one would you run? (I guess you're aware of the standard results saying a non-stupid AI must follow some one-dimensional utility function, etc.)

Comment author: ArisKatsaris 10 January 2014 04:06:06PM 0 points [-]

If you were offered a bunch of AIs with equivalent power, but following different mixtures of your moral and non-moral preferences, which one would you run?

I guess whatever ratio of my moral and non-moral preferences best represents their effect on my volition.

Comment author: TheOtherDave 10 January 2014 02:31:40PM 0 points [-]

My related but different thoughts here. In particular, I don't agree that emotions like moral outrage and approval are impersonal, though I agree that we often justify those emotions using impersonal language and beliefs.

Comment author: ArisKatsaris 10 January 2014 03:34:14PM 0 points [-]

I didn't say that moral outrage and approval are impersonal. Obviously nothing that a person does can truly be "impersonal". But it may be an attempt at impersonality.

The attempt itself provides a direction that significantly differentiates between moral preferences and non-moral preferences.

Comment author: TheOtherDave 10 January 2014 06:21:21PM 0 points [-]

I didn't mean some idealized humanly-unrealizable notion of impersonality, I meant the thing we ordinarily use "impersonal" to mean when talking about what humans do.

Comment author: Kawoomba 10 January 2014 01:48:38PM *  0 points [-]

IMO the only sane reaction to that argument is to unify the concepts of "wishes" and "morality" into a single concept, which you could call "preference" or "morality" or "utility function", and just switch to using it exclusively. I've made that switch so long ago that I've forgotten how to think otherwise.

Ditto.

Cousin Itt, 'tis a hairy topic, so you're uniquely "suited" to offer strands of insights:

For all the supposedly hard and confusing concepts out there, few have such an obvious answer as the supposed dichotomy between "morality" and "utility function". This in itself is troubling, as too-easy-to-come-by answers trigger the suspicion that I myself am subject to some sort of cognitive error.

Many people I deem to be quite smart would disagree with you and I, on a question whose answer is pretty much inherent in the definition of the term "utility function" encompassing preferences of any kind, leaving no space for some holier-than-thou universal (whether human-universal, or "optimal", or "to be aspired to", or "neurotypical", or whatever other tortured notions I've had to read) moral preferences which are somehow separate.

Why do you reckon that other (or otherwise?) smart people come to different conclusions on this?

Comment author: cousin_it 10 January 2014 02:31:22PM *  1 point [-]

I guess they have strong intuitions saying that objective morality must exist, and aren't used to solving or dismissing philosophical problems by asking "what would be useful for building FAI?" From most other perspectives, the question does look open.

Comment author: TheAncientGeek 10 January 2014 02:32:32PM 0 points [-]

Moral preferences don't have to be separate to be disinct, they can be a subset. "Morality is either all your prefences, or none of your prefernces" is a false dichotomy.

Comment author: Kawoomba 10 January 2014 02:35:35PM *  1 point [-]

Edit: Of course you can choose to call a subset of your preferences "moral", but why would that make them "special", or more worthy of consideration than any other "non-moral" preferences of comparative weight?

Comment author: ArisKatsaris 10 January 2014 03:32:28PM 0 points [-]

The "moral" subset of people's preferences has certain elements that differentiate it like e.g. an attempt at universalization.

Comment author: Kawoomba 10 January 2014 03:41:36PM 0 points [-]

Attempt at universalization, isn't that a euphemism for proselytizing?

Why would [an agent whose preferences do not much intersect with the "moral" preferences of some group of agents] consider such attempts at universalization any different from other attempts of other-optimizing, which is generally a hostile act to be defended against?

Comment author: ArisKatsaris 10 January 2014 03:57:57PM *  1 point [-]

Attempt at universalization, isn't that a euphemism for proselytizing?

No, people attempt to 'proselytize' their non-moral preferences too. If I attempt to share my love of My Little Pony, that doesn't mean I consider it a need for you to also love it. Even if I preferred it for you share my love of it, it would still not be a moral obligation on your part.

By universalization I didn't mean any action done after the adoption of the moral preference in question, I meant the criterion that serves to label it as a 'moral injuction' in the first place. If your brain doesn't register it as an instruction defensible by something other that your personal preferences, if it doesn't register it as a universal principle, it doesn't register as a moral instruction in the first place.

Comment author: Kawoomba 10 January 2014 04:03:24PM *  0 points [-]

What do you mean by "universal"? For any such "universally morally correct preference", what about the potentially infinite number of other agents not sharing it? Please explain.

Comment author: TheAncientGeek 10 January 2014 03:27:17PM *  0 points [-]

The key issue is that, whilst morality is not tautologously the same as preferences, a morally right action is, tautologously, what you should do.

So it is difficult to see on what grounds Mark can object to the FAIs wishes: if it tells him something is mortally right that is what he should do. And he can't have his own separate morality, because the idea is incoherent.

Comment author: ArisKatsaris 10 January 2014 03:44:58PM *  0 points [-]

So it is difficult to see on what grounds Mark can object to the FAIs wishes: if it tells him something is mortally right that is what he should do.

A distinction to be made: Mark can wish differently than the AI wishes, Mark can't morally object to the AI's wishes (if the AI follows morality).

Exactly because morality is not the same as preferences.

Comment author: Kawoomba 10 January 2014 03:35:03PM *  0 points [-]

You can call a subset of your preferences moral, that's fine. Say, eating chocolate icecream, or helping a starving child. Let's take a randomly chosen "morally right action" A.

That, given your second paragraph, would have to be a preference which, what, maximizes Mark's utility, regardless of what the rest of his utility function actually looks like?

It seems to be trivial to construct a utility function (given any such action A) such as that doing A does not maximize said utility function. Give Mark a such a utility function and you got yourself a reductio ad absurdum.

So, if you define a subset of preferences named "morally right" thus that any such action needs to maximize (edit: or even 'not minimize') an arbitrary utility function, then obviously that subset is empty.

Comment author: TheAncientGeek 10 January 2014 04:00:46PM 0 points [-]

That, given your second paragraph, would have to be a preference which, what, maximizes Mark's utility, regardless of what the rest of his utility function actually looks like?

If Mark is capable of acting morally, he would have a preference for moral action which is strong enough to override other preferences. However,t hat is not really the point. Even if he is too weak-willed to do what the FAI says, he has no grounds to object to the FAI.

It seems to be trivial to construct a utility function (given any such action A) such as that doing A does not maximize said utility function. Give Mark a such a utility function and you got yourself a reductio ad absurdum.

I can't see how that amount to more than the observation that not every agetn is capable of acting morally. Ho hum.

So, if you define a subset of preferences named "morally right" thus that any such action needs to maximize (edit: or even 'not minimize') an arbitrary utility function, then obviously that subset is empty.

I don't see why. An agent should want to do what is morally right, but that doesn't mean an agent would want to. Their utility funciton might not allow them. But how could they object to be told what is right? The fault, surely, lies in themselves.

Comment author: Kawoomba 10 January 2014 04:14:45PM 0 points [-]

An agent should want to do what is morally right, but that doesn't mean an agent would want to. Their utility funciton might not allow them. But how could they object to be told what is right? The fault, surely, lies in themselves.

They can object because their preferences are defined by their utility function, full stop. That's it. They are not "at fault", or "in error", for not adopting some other preferences that some other agents deem to be "morally correct". They are following their programming, as you follow yours. Different groups of agents share different parts of their preferences, think Venn diagram.

If the oracle tells you "this action maximizes your own utility function, you cannot understand how", then yes the agent should follow the advice.

If the oracle told an agent "do this, it is morally right", the non-confused agent would ask "do you mean it maximizes my own utility function?". If yes, "thanks, I'll do that", if no "go eff yourself!".

You can call an agent "incapable of acting morally" because you don't like what it's doing, it needn't care. It might just as well call you "incapable of acting morally" if your circles of supposedly "morally correct actions" don't intersect.

Comment author: TheOtherDave 10 January 2014 02:21:29PM 0 points [-]

I can't speak for cousin_it, natch, but for my own part I think it has to do with mutually exclusive preferences vs orthogonal/mutually reinforcing preferences. Using moral language is a way of framing a preference as mutually exclusive with other preferences.

That is... if you want A and I want B, and I believe the larger system allows (Kawoomba gets A AND Dave gets B), I'm more likely to talk about our individual preferences. If I don't think that's possible, I'm more likely to use universal language ("moral," "optimal," "right," etc.), in order to signal that there's a conflict to be resolved. (Well, assuming I'm being honest.)

For example, "You like chocolate, I like vanilla" does not signal a conflict; "Chocolate is wrong, vanilla is right" does.

Comment author: TheAncientGeek 10 January 2014 02:56:40PM *  0 points [-]

Why stop at connotation and signalling? If there is a non-empty set of preferences whose satistfaction is inclined to lead to conflict, and a non-empty set of preferences that can be satisfied withotu conflict, then "morally relevant prefernece" can denote the members of the first set...which is not idenitcal to the set of all preferences.

Comment author: Kawoomba 10 January 2014 03:52:22PM 0 points [-]

For any such preference, you can immediately provide a utility function such that the corresponding agent would be very unhappy about that preference, and would give its life to prevent it.

Or do you mean "a set of preferences the implementation of which would on balance benefit the largest amount of agents the most"? That would change as the set of agents changes, so does the "correct" morality change too, then?

Also, why should I or anyone else particular care about about such preferences (however you define them), especially as the "on average" doesn't benefit me? Is it because evolutionary speaking, that's how what evolved? What our mirror neurons lead us towards? Wouldn't that just be a case of the naturalistic fallacy?

Comment author: TheAncientGeek 10 January 2014 04:13:01PM 0 points [-]

For any such preference, you can immediately provide a utility function such that the corresponding agent would be very unhappy about that preference

Sure. So what? Kids don't like teachers and criminals don't like the police..but they can't object to them, because "entitiy X is stopping from doing bad things and making me do good things" is no (rational, adult) objection.

Also, why should I or anyone else particular care about about such preferences (however you define them), especially as the "on average" doesn't benefit me?

If being moral increases your utility, it increases your utility -- what other sense of "benefitting me" is there?

Comment author: blacktrance 10 January 2014 04:26:19PM 0 points [-]

If being moral increases your utility, it increases your utility -- what other sense of "benefitting me" is there?

If utility is the satisfaction of preferences, and you can have preferences that don't benefit you (such as doing heroin), increasing your utility doesn't necessarily benefit you.

Comment author: TheAncientGeek 10 January 2014 04:37:30PM 0 points [-]

If you can get utility out of paperclips, why can't you get it out of heorin? You're surely not saying that there is some sort of Objective utility that everyone ought to have in their UF's?

Comment author: Kawoomba 10 January 2014 04:17:53PM 0 points [-]

Kids don't like teachers and criminals don't like the police..but they can't object to them (...)

(Kids - teachers), (criminals - police), so is "morally correct" defined by the most powerful agents, then?

If being moral increases your utility (...)

And if being moral (whatever it may mean) does not?

Comment author: TheAncientGeek 10 January 2014 04:47:56PM *  -1 points [-]

(Kids - teachers), (criminals - police), so is "morally correct" defined by the most powerful agents, then?

Adult, rational objections are objections that other agents might feel impelled to do somehting about, and so are not just based on "I don't like it"."I don't like it" is no objectio to "you should do your homework", etc.

If being moral increases your utility (...)

And if being moral (whatever it may mean) does not?

Then you would belong to the set of Immoral Agents, AKA Bad People.

Comment author: TheOtherDave 10 January 2014 02:08:38PM 1 point [-]

For my own part: denotationally, yes, I would understand "Do you prefer (that Dave eat) chocolate or vanilla ice cream?" and "Do you consider (Dave eating) chocolate ice cream or vanilla as the morally superior flavor for (Dave eating) ice cream?" as asking the same question.

Connotationally, of course, the latter has all kinds of (mostly ill-defined) baggage the former doesn't.

Comment author: TheAncientGeek 10 January 2014 04:26:18PM 0 points [-]

Are you using a meaning of "morality" that is distinct from "preferences"? You bet.