Wei_Dai comments on Where do selfish values come from? - Less Wrong

27 Post author: Wei_Dai 18 November 2011 11:52PM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (57)

You are viewing a single comment's thread. Show more comments above.

Comment author: Wei_Dai 19 November 2011 07:23:07AM *  7 points [-]

It seems that in this post, by "selfish" you mean something like "not updateless" or "not caring about counterfactuals".

By "selfish" I mean how each human (apparently) cares about himself more than others, which needs an explanation because there can't be a description of himself embedded in his brain at birth. "Not updateless" is meant to be a proposed explanation, not a definition of "selfish".

A meaning closer to usual sense of the word would be, "caring about welfare of a particular individual" (including counterfactual instances of that individual, etc.), which seems perfectly amenable to being packaged as a reflectively consistent agent (that is not the individual in question) with world-determined utility function.

No, that's not the meaning I had in mind.

(A reference to usage in Stuart's paper maybe? I didn't follow it.)

This post isn't related to his paper, except that it made me think about selfishness and how it relates to AIXI and UDT.

Comment author: Vladimir_Nesov 19 November 2011 01:08:30PM *  6 points [-]

By "selfish" I mean how each human (apparently) cares about himself more than others, which needs an explanation because there can't be a description of himself embedded in his brain at birth.

Pointing at self is possible, which looks like a reasonable description of self, referring to all the details of a particular person. That is, interpretation of individual's goal representation depends on the fact that the valued individual is collocated with the individual-as-agent.

Just as how a file offset value stored in memory of my computer won't be referring to the same data if used on (moved to) your computer that has different files; its usefulness depends on the fact that it's kept on the same computer; and it will continue to refer to same data if we move the whole computer around.

No, that's not the meaning I had in mind.

Now I'm confused again, as I don't see how these senses (one I suggested and one you explained in parent comment) differ, other than on the point of caring vs. not caring about counterfactual versions of same individual. You said, "each human (apparently) cares about himself more than others, which needs an explanation", and it reads to me as asking how can humans have the individual-focused utility I suggested, that you then characterized as not the meaning you had in mind...

Comment author: [deleted] 19 November 2011 07:53:33PM 3 points [-]

there can't be a description of himself embedded in his brain at birth.

Why not, or what do you mean by this? Common sense suggests that we do know ourselves from others at a very low, instinctive level.

Comment author: Vladimir_Nesov 19 November 2011 09:38:46PM *  3 points [-]

I expect Wei's intuition is that knowing self means having an axiomatic definition of (something sufficiently similar to) self, so that it can be reasoned about for decision-theoretic purposes. But if we look at an axiomatic definition as merely some structure that is in known relation to the structure it defines, then your brain state in the past is just as good, and the latter can be observed in many ways, including through memory, accounts of own behavior, etc., and theoretically to any level of detail.

(Knowing self "at a low, instinctive level" doesn't in itself meet the requirement of having access to a detailed description, but is sufficient to point to one.)

Comment author: torekp 25 November 2011 04:44:50PM 2 points [-]

I think there's enough science on the subject - here's the first paper <pdf> I could find with a quick Google - to sketch out an approximate answer to the question of how self-care arises in an individual life. The infant first needs to form the concept of a person (what Bischof calls self-objectification), loosely speaking a being with both a body and a mind. This concept can be applied to both self and others. Then, depending on its level of emotional contagion (likelihood of feeling similarly to others when observing their emotions) it will learn, through sophisticated operant conditioning, self-concern and other-concern at different rates.

Since the typical human degree of emotional contagion is less than unity, we tend to be selfish to some degree. I'm using the word "selfish" just as you've indicated.

Comment author: snarles 24 November 2011 11:57:52AM 1 point [-]

Just as altruism can be related to trust, selfishness can be related to distrust.

An agent which has a high prior belief in the existence of deceptive adversaries would exhibit "selfish" behaviors.

Comment author: Stuart_Armstrong 19 November 2011 12:58:19PM 1 point [-]

No, that's not the meaning I had in mind.

What is your meaning then? What would you call "caring about the welfare of a particular individual (that happens to be myself)"?

Comment author: Wei_Dai 19 November 2011 07:11:02PM 2 points [-]

Ok, I do mean:

caring about the welfare of a particular individual (that happens to be myself)

but I don't mean:

caring about welfare of a particular individual

(i.e., without the part in parenthesis) Does that clear it up?

Comment author: Vladimir_Nesov 19 November 2011 09:25:00PM *  1 point [-]

Do you mean that the agent itself must be the person it cares about? What if the agent is carried in a backpack (of the person in question), or works over the Internet?

What if the selfish agent that cares about itself writes an AI that cares about the agent, giving this AI more optimization power, since they share the same goal?

Comment author: Stuart_Armstrong 21 November 2011 10:33:41AM 0 points [-]

Ah, there was a slight confusion on my part. So if I'm reading this correctly you define formally selfish to mean... selfish. :-)