Have you read Mark Twain's "What is Man"? If I recall correctly, there he lays out his argument that man is already always selfish. For example, we do good deeds ultimately for our own comfort, because they make us feel good. (If I recall rightly, he also makes the, to me rather more interesting, point that people are born happy or sad rather than are made happy or sad by specific, supposedly uplifting or depressing, thoughts. That seems to anticipate the modern pharmaceutical approach to mood regulation.)
For my part, I think that "selfish&q...
Did Hopefully Anonymous figure this out and stop expending effort commenting or posting on his anonymous blog?
So, it seems that Eliezer's working definition of an intelligent person is "someone who agrees with me".
I must point out that "whenever they want to be nice to someone" entails a desire to be nice to someone. Your very phrase defines it as being in their interests to be nice to someone. Rationalization isn't even necessary here. You wanted to do something - you did it. Selfishness isn't that complicated.
My guess would be that this individual had read Atlas Shrugged and hadn't fully understood what selfish meant in the context. Ayn Rand was setting out to redefine the word, not to glorify the "old" meaning.
I think that people are born selfish. This is based on the fact that, if I grew up in an environment that didn't compel me to be either selfish or selfless, I'd probably be selfish. Babies are selfish. I value those creatures that I was convinced to value. Slave owners taught their children that it's okay to beat slaves, and the children were happy to comply. Now most people disregard the pain of food animals because they can get away with it.
Of course, some of my actions are genuinely altruistic. I chose to give up meat, although this brings me little tangible benefit. (It does get me out of some accusations of hypocricy.) One reason why I let myself become like this is that in human society, being nice is a habit that keeps my ass from getting kicked. And it needs to be a habit, because I'm not smart enough to delude everybody that I care, when I actually see them all as obstacles.
If I somehow become so powerful that I no longer depend on anyone, and noting they do can harm me, I will probably quickly become corrupted by my power.
But I agree that now, I can't be considered purely selfish.
"So, it seems that Eliezer's working definition of an intelligent person is "someone who agrees with me"."
My definition of an intelligent person is slowly becoming "someone who agrees with Eliezer", so that's all right. Plus, the guy showed ability to revise a strongly held belief.
Read the comments at Hopefully Anonymous' most recent post. He explains why he has been inactive.
I want you to be altruistic, Eliezer. That's partly because I think you're intelligent. I would prefer if some people were more selfish though.
I choose living in comfort for one more year. There are things I might die for, but I don't know what exactly. Perhaps to spite someone. If other people knew that I had the chance to save the world and were going to punish me for failing to do so, I might not risk their wrath. I also choose the stanger getting tortured, ...
"My definition of an intelligent person is slowly becoming 'someone who agrees with Eliezer', so that's all right."
That's not in the spirit of this blog. Status is the enemy, only facts are important.
"That's not in the spirit of this blog. Status is the enemy, only facts are important."
See? Another smart man agrees with Eliezer. That's what I'm talking about.
Taking a cue from some earlier writing by Eli, I suppose one way to give ethical systems a functional test is to imagine having access to a genie. An altruist might ask the genie to maximize the amount of happiness in the universe or something like that, in which case the genie might create a huge number of wireheads. This seems to me like a bad outcome, and would likely be seen as a bad outcome by the altruist who made the request of the genie. A selfish person might say to the genie "create the scenario I most want/approve of." Then it would be...
"Do not seek to follow in the footsteps of the wise, seek what they sought." -- Nanzan Daishi, quoted by Matsuo Basho.
Stephen: the altruist can ask the Genie the same thing as the selfish person. In some sense, though, I think these sorts of wishes are 'cheating,' because you are shifting the computational/formalization burden from the wisher to the wishee. (Sorry for the thread derail.)
"An altruist might ask the genie to maximize the amount of happiness in the universe or something like that, in which case the genie might create a huge number of wireheads. This seems to me like a bad outcome, and would likely be seen as a bad outcome by the altruist who made the request of the genie."
Eh? An altruist would voluntarily summon disaster upon the world?
By the way, I have some questions about wireheading. What is it, really? Why is it so repulsive? Is it really so bad? If, when you imagine your brain rewired, you envision something t...
Humans seem to gain social status by persuading others to agree with them. This is one of the reasons we resist being persuaded by good arguments. So a selfish person who wanted social status could want you to be selfish also in order to gain social status, showing our dominance via the submission of others.
It seems to me that this discussion is somewhat misleading. Each one of us members of the Homo Sapiens species operates to pursue his/her own interests, as dictated by her/his genetic code. But in order to do so, we have got to cooperate with some of our fellow human beings. Each and every one of our actions is in some ways a combination of these two typical aspects of our behavior. There is no such thing as a totally selfish or a totally unselfish behavior/action/activity, much less can we talk of a totally selfish or a totally unselfish person -- only ex...
In my above comment at 5:49 AM, the sentence "Whether we chose to declare ourselves selfish or unselfishness" should actually read: "Whether we chose to declare ourselves selfish or unselfish"
I said, "If you're genuinely selfish, then why do you want me to be selfish too? Doesn't that make you concerned for my welfare? Shouldn't you be trying to persuade me to be more altruistic, so you can exploit me?"
The objection you press against your interlocutor was anticipated by Max Stirner, the renowned champion of egoism, who replied as follows:
Do I write out of love to men? No, I write because I want to procure for my thoughts an existence in the world; and, even if I foresaw that these thoughts would deprive you of your rest and your pea...
Selfishness - or to avoid confusion, let's call the concept 'self-interest' - takes on rather a different appearance when it's realized that the 'self' is not something necessarily limited to the boundaries of the physical form that embodies the distinction.
To the degree that we identify with and value the rest of humanity, sacrificing one's own existence to preserve the rest of humanity can be in the self-interest. To the degree that we don't, or that we negatively value the rest of humanity, that action can be against self-interest. If we disliked humanity enough, we'd choose to destroy it even if it cost us our own lives (which presumably we'd value) in the process.
So, it seems that Eliezer's working definition of an intelligent person is "someone who agrees with me".
Communities form because they repress the incompatible. Particularly on the Internet, where it's easy to restrict who can participate, people tend to agree not because they persuade one another but because they seek out and associate with like-minded people.
Obviously Eliezer thinks that the people who agree with the arguments that convince him are intelligent. Valuing people who can show your cherished arguments to be wrong is very nearly a post-human trait - it is extraordinarily rare among humans, and even then unevenly manifested.
If we posit an ideal world where every person has perfect and complete knowledge, and the discipline and self-control to act consistently on that knowledge, it's possible we can equate the most self-interested act with the most ethical.
Until we have that ideal world, to posit that people always simply do what they want to do anyway, and rationalize their behavior to their philosophy of life, is to engage in a bit of the same rationalization when we conclude that "if only they knew as much as I know, they would do what I think they should do."
Too ...
Are providing answers to questions like "Would you do incredible thing X if condition Y was true" really necessary if thing X is something neither person would likely ever be able to do and condition Y is simply never going to happen? It seems easy to construct impossible moral challenges to oppose a particular belief, but why should beliefs be built around impossible moral edge cases? Shouldn't a person be able to develop a rational set of beliefs that do fail under extreme moral cases, but at the same time still hold a perfectly strong and not contradictory position?
Obviously Eliezer thinks that the people who agree with the arguments that convince him are intelligent. Valuing people who can show your cherished arguments to be wrong is very nearly a post-human trait - it is extraordinarily rare among humans, and even then unevenly manifested.
On the other hand, if we are truly dedicated to overcoming bias, then we should value such people even more highly than those whom we can convince to question or abandon their cherished (but wrong) arguments/beliefs.
The problem is figuring out who those people are.
But it's very di...
It's not that the one agreed with me and declared himself no longer selfish, but that he showed nonzero reactivity in the face of an unexpected argument, a rare thing. Further conversation (not shown) did seem to show that he was thinking about it. You don't see, say, Caledonian ever updating his views, or showing nonzero dependency between what I say and his ability to comment negatively on every post.
Ah, but I have yet to be confronted with an argument that would cause me to update my views.
So now the problem expands: if two people disagree about the worth of an argument, what criteria do we use to choose between them? What if we ARE one of the two people? How do we use our own judgement to evaluate our own judgement?
Simply, we can't. At best, we can look for simple errors that we presume we can objectively determine, and leave the subtle issues to other systems.
I don't see you admitting you were wrong in the previous threads, Eliezer. Should I interpret that as your unwillingness to admit to error, or that you're so much smarter than me that I can't even comprehend how you're actually correct?
"Selfish" in the negative sense means not just pursuing one's own interests, but doing so heedless of the harm one's actions may be causing others. I don't think there are many proponents of "selfishness" in this sense.
There are people that are "selfless" in the sense that they not only don't act according to their direct self-interest, they even abandon their own concepts of true and false, right and wrong, trusting some external authority to make these judgments for them. Religious, political, whatever. People who praise selfishness are generally contrasting it with this kind of selflessness.
My understand is that the philosophy of rational self-interest, as forwarded by the Objectivists, contains a moral system founded first on the pursuit of maintaining a high degree of "conceptual" volitional consciousness and freedom as a human being. Anything that robs one's life or robs one's essential humanity is opposed to that value. The Objectivist favor of capitalism stems from a belief that capitalism is a system that does much to preserve this value (the essential freedom and humanity of individuals). Objectivists are classical libertaria...
like pablo I have considered whether arguing for in my case altruism/utilitarianism is always altruistic and thought "well probably" - but I dont analyse it much because in the end I don't know if it would matter - it seems I do what I do because 'that is what I am', more than 'that is what I think is right'. I guess it works the other way too eh.
Michael Sullivan:
That's an exceptionally clear exegesis. Thanks!
Pablo Stafforini:
The words of Max Stirner (with whom I am admittedly unfamiliar) that you quote seem to me like so much bluster and semantic question-begging.
Do I write out of love to men? No, I write because I want to procure for my thoughts an existence in the world; and, even if I foresaw that these thoughts would deprive you of your rest and your peace, even if I saw the bloodiest wars blah blah blah
He sings not out of love for the hearer, but because he loves to sing and the hearer is us...
Daniel, you can read Stirner's book here. It's not really "macho", that would be more Ragnar Redbeard's "Might Makes Right".
Among the things Stirner writes about is freedom of expression. He does not care for what the state or the church say he may write, because he takes such freedom for himself (many unauthorized printings of his book were made). He does not respect the holy but instead regards taboos against blasphemy as attempted restrictions on him that he will violate. For people who say that he ought not to speak of certain thing...
Eliezer's descriptions of his intellectual history and past mistakes are very convincing positive signals to me.
I agree, but have a nagging doubt. When I read years-old writings where he makes some of those mistakes he sounds about as knowledgeable, just as smart, just as honest, and just as sure he's right as he is now in his new beliefs.
Although I was convinced by very few of those older mistakes (before I searched and found retractions) that could just as easily mean new arguments got super persuasive rather than super accurate.
His writings have convinc...
PS. I know arguments should be judged based on their own worth and not who made them, but there are other factors.
Sometimes people change their minds because new evidence and novel arguments have forced them to re-evaluate their positions.
Other times, people have never thought through even the obvious arguments, or are convinced easily by weak data and unimpressive theses, so they shift from one point to another.
If someone changes their positions frequently, AND they're very confident in their positions, that's a bad sign.
Eliezer: “Other mischievous questions to ask self-proclaimed Selfishes: "Would you sacrifice your own life to save the entire human species?" (If they notice that their own life is strictly included within the human species, you can specify that they can choose between dying immediately to save the Earth, or living in comfort for one more year and then dying along with Earth.)”
What If the person saving the entire human species wanted to commit suicide anyway or if he/she has this dream of heroism and hopes to become immortal and be rewarded in an afterlife. An apparently selfless deed can be a very selfish one.
The question Eliezer raises is the first problem any religious person has to face once he abandons the god thesis, i.e. why should I be good now? The answer, I believe, is that you cannot act contrary to your genetic nature. Our brains are wired (or have modules in Pinker terms) for various forms of altruism, for group survival reasons probably. I therefore can’t easily commit acts against my genetic nature, even if intellectually I can see they are in my best interests. (As Eliezer has already recognised this is why AI or uploaded personalities are so dan...
If someone changes their positions frequently, AND they're very confident in their positions, that's a bad sign.
I don't think he changes his mind too frequently, and being overly confident at the time of now-abandoned positions isn't unusual. My point was that in Eliezer's case a knowledgeable, smart, honest, and self-certain argument doesn't imply strong evidence of truth, because those qualities appear in arguments that turned out false.
To be honest I think I was hoping someone would leap to his defense and crush my argument, giving me permission to be...
ChrisA, why are you intellectually convinced you should be selfish? Rationality doesn't demand any particular goals. A genuinely altruistic person, if uploaded, would overwrite the "brain code" (a bad analogy; evolved tendencies aren't deterministic "code") that promotes selfishness.
Recovering,
While I wish I had something reassuring to say on this subject, you should probably be quite disturbed if you find my work from 1997 sounding as persuasive as my work from 2007.
Nick
My response is, evolution! Let's say a genuinely (what ever that means) altruistic entity exists. He then is uploaded. He then observed that not all entities are fully altruistic, in other words they will want to take resources from others. In any contest over resources this puts the altruistic entity at a disadvantage (he is spending resources helping others that he could use to defend himself). With potentially mega intelligent entities any weakness is serious. He realises that very quickly he will be eliminated if he doesn't fix this weakness. He ei...
While I wish I had something reassuring to say on this subject, you should probably be quite disturbed if you find my work from 1997 sounding as persuasive as my work from 2007
But I said...
Although I was convinced by very few of those older mistakes (before I searched and found retractions) that could just as easily mean new arguments got super persuasive rather than super accurate.
All my comments mean in practice is that even though once I study and investigate your (new) arguments they nearly always seem to me to be right, I won't let myself start lazily...
Fair enough, Recovering. My own point is that:
When I read years-old writings where he makes some of those mistakes he sounds about as knowledgeable, just as smart, just as honest, and just as sure he's right as he is now in his new beliefs.
Then those factors aren't very good discriminators of truth, are they? It's not just "improper" to take them into account, it actually doesn't work.
In whatever facets I sounded about as "knowledgeable", "smart", "honest", or "self-assured" then as now, you might take th...
In whatever facets I sounded about as "knowledgeable", "smart", "honest", or "self-assured" then as now, you might take these facets into account in deciding whether someone's arguments are worth your time to read, but you shouldn't take them into account in deciding whether the person is right.
Agreed. Having said that, I do find those facets to correlate with truth, but the correlation flattens out for high values. Besides, the first two would be hard for me to judge well between your 1997 and 2007 selves, for obvio...
When I say that I am selfish, I mean to express that I think the best model for "altruism" is that of a good consumed much like any other. I consume it for my personal enjoyment, not in proportion to the benefit received by the recipient. And, ceteris paribus, with a declining marginal utility as quantity consumed increases.
In my eyes, a "true" altruist would value the 1000th meal provided to a starving third-worlder on par with the 1st one provided, given that the beneficiaries valued the meals similarly. Nobody behaves that way. Altru...
Mark, altruists have to deal with their costs too.
It's possible for an altruist to value the thousandth altruistic meal as much as the first, but as his resources shrink the value of the alternatives rises. If I provide meals for a hundred thousand starving people and then I have nothing left and I become a starving person myself, that isn't good. At some point I want to keep enough capital to maintain my continuing ability to feed starving people.
I'm not claiming that it's true that no altruist experiences diminishing returns, or even that there is an altruist who doesn't experience diminishing returns. But the behavior doesn't prove that there couldn't be, and so this isn't a killing blow.
J Thomas,
A non-diminishing-returns altruist would hit a point where the utility of spending a marginal resource on a "selfish" purpose dips below the best use of that resource for an "altruistic" purpose. Every single marginal resource after that should go toward altruistic purposes as well. Why? Because for anyone with non-astronomical resources, there is effectively an endless supply of altruistic options that all provide effectively the same degree of benefit to the recipient. The non-diminishing-returns altruist would increase altru...
As a selfish prizefighter, I want to beat my opponent. If I was an altruist instead, I don't think I'd be able to win one fight. Because I am in fact selfish, fighting an opponent who is an altruist would not do much for my self-esteem. Only in fighting better fighters than I am do I learn, not by fighting someone inferior. If a superior fighter does not do his best in a given match with me for some reason, I cannot objectively pretend to be better than him just because I won once. It benefits me to beat him when he's at his best. I like to share my knowle...
Re: It looks to me like when people espouse a philosophy of selfishness, it has no effect on their behavior
This is almost certainly not true of conscious genetic selfishness. Such individuals can be expected to engage in various rare and unusual activities - such as donating to sperm banks.
There's all kinds of incoherency with the idea of "selfishness" being defined as "acting in your own interest." What is your own interest? Can't you define it circularly as being anything you desire to do? Being a "selfish person" doesn't necessarily make sense by that definition.
Maybe it's better to go the opposite direction. Selfishness is indifference to the desires and well-being of others. Whatever else you may be doing (let's drop the question of whether it's "self-interested" or not), it's more important ...
I'm relieved that at least a few people mentioned Stirner. The selfishness EY portrays is not representative of selfishness in the sense any of the literature of Philosophical Egoism of which I am aware. His scenario doesn't even correspond to what a Randian might believe about selfishness.
Max Stirner is the best and most intellectually consistent egoist I'm aware of. Less accomplished but more polemical writers in the same vein include John L. Walker, Benjamin Tucker, John Badcock, Dora Marsden (for a period of time), and Sid Parker.
Max Stirner's "...
I've always felt that both "selfishness" and "altruism" were results of the Fundamental Attribution Error. Some actions are deemed "selfish" according to society's mean value set, others are deemed "altruistic" or "selfless". Personally, I'm more interested in the chain of events that ultimately lead up to an action being performed and in the chain of events that occurred as a result of it than I am in applying labels of dubious and limited utility to things.
"Would you steal a thousand dollars from Bill Gates if you could be guaranteed that neither he nor anyone else would ever find out about it?"
Without making a case for or against my own altruism, I'd definitely answer yes to this question, for two reasons, which are really one reason:
1) From a utilitarian standpoint, a thousand dollars benefits me much more than losing a thousand dollars harms Bill Gates. Much, much, much more.
2) I am destitute and in grave need of money
Is this fair reasoning? I'd also say the monetary status of the individual ...
"If you're genuinely selfish, then why do you want me to be selfish too? Doesn't that make you concerned for my welfare? Shouldn't you be trying to persuade me to be more altruistic, so you can exploit me?"
You're conflating selfishness with vulgar egoism. Suppose your well-being makes me happy, and I believe that making you selfish will make you happier. Then convincing you to be selfish is the self-interested thing to do. If I tried to convince you to be more altruistic (in the self-sacrificing sense, not in the benevolent sense) so I could...
If I didn't know that someone was going to be tortured, I would rather stub my toe, and I do not claim to be selfish. Otherwise I am not really sure how to interpret the question.
Many big-L Libertarians I've met - along with those who consider themselves to be trench-fighters for Ayn Rand-ian Objectivism - seem to want to conflate "selfishness" with "enlightened self-interest" for the positive connotations of the latter... yet their rationale for various big-L proposals (such as "let's turn over national security to corporations, who will certainly never abuse the power to force decisions upon people") tends to be of the extremely rosy, happy death spiral, declare-anything-that-doesn't-fit-an-"ext...
I think this whole egoism vs. altruism debate is too much black and white thinking.
What about evolutionary ethics? Doing what is good for your genes and the survival of the species.
Therefore, one would put self-interest over the interest of other people, but not totally disregard the latter. Also some people like your children would be more important than others. The ranking would look something like this:
Selfishness is the default, but can be overridden if the bad consequences times...
I used to think about self-proclaimed selfishness in this way all the time, but now I think I "get" the doctrine of selfishness. For instance, the answer to the first mischevous question is actually "because I get personal happiness from preaching selfishness, and that vastly overshadows any happiness I could instead get from a single person's altruism". The answer to own life vs. human species is "because a guilt-ridden life wouldn't maximise my personal happiness", and so on.
When we weigh options in our mind, we pick the one that yields the cocktail of chemicals/neurotransmitters that induce the strongest positive response in our reward center. Or rather, the cocktail of chemicals/neurotransmitters that elicits the strongest positive response Is able to pass its signals through to the motor neurons.
A desire to be moral, a desire to avoid pain, a desire to protect kin, all release chemicals.
Seen in this light, the phrase "everything one does is selfish" appears to reduce to "all choices are weighed through one's...
The one replied: "Well, if you become selfish, then you'll realize that it's in your rational self-interest to play a productive role in the economy, instead of, for example, passing laws that infringe on my private property."
I get that this isn't the point of the post but... what? No, just no. If I'm selfish I'm going to pass laws that don't infringe on MY private property, why should I care about yours? Indeed, I'll just go on with using any shred of political influence I have to make sure your taxes end up paying my expenses and thus increas...
My answer to your first mischievous question tends to be: "if you (also) identify as selfish, this will make you more predictable and thereby more trustworthy".
Don't give two shits about one guy's contribution to the economy. My selfish incentives are purely local.
Besides, it's always better to cooperate with a rich guy than exploit a poor one.
But of course there's no need to convince you to be something you already are
"f they say that they'd be emotionally disturbed by knowing, specify that they won't know about the torture."
Couldn't one argue that having preferences about things u assume u dont know, wouldn't affect your actions?
When I'm deciding on an actual action, I can only take into account things I know, and nothing else?
So the preference of the case where I would never know about the person being tortured couldn't affect my actions, so in that sense doesn't matter?
You can't expect a truly selfish agent to answer any of these questions honestly. Admitting you would steal in certain circumstances would cause them reputational damage. Most of us have programming in place to prevent us from incriminating ourselves morally.
Once upon a time, I met someone who proclaimed himself to be purely selfish, and told me that I should be purely selfish as well. I was feeling mischievous(*) that day, so I said, "I've observed that with most religious people, at least the ones I meet, it doesn't matter much what their religion says, because whatever they want to do, they can find a religious reason for it. Their religion says they should stone unbelievers, but they want to be nice to people, so they find a religious justification for that instead. It looks to me like when people espouse a philosophy of selfishness, it has no effect on their behavior, because whenever they want to be nice to people, they can rationalize it in selfish terms."
And the one said, "I don't think that's true."
I said, "If you're genuinely selfish, then why do you want me to be selfish too? Doesn't that make you concerned for my welfare? Shouldn't you be trying to persuade me to be more altruistic, so you can exploit me?"
The one replied: "Well, if you become selfish, then you'll realize that it's in your rational self-interest to play a productive role in the economy, instead of, for example, passing laws that infringe on my private property."
And I said, "But I'm a small-L libertarian already, so I'm not going to support those laws. And since I conceive of myself as an altruist, I've taken a job that I expect to benefit a lot of people, including you, instead of a job that pays more. Would you really benefit more from me if I became selfish? Besides, is trying to persuade me to be selfish the most selfish thing you could be doing? Aren't there other things you could do with your time that would bring much more direct benefits? But what I really want to know is this: Did you start out by thinking that you wanted to be selfish, and then decide this was the most selfish thing you could possibly do? Or did you start out by wanting to convert others to selfishness, then look for ways to rationalize that as self-benefiting?"
And the one said, "You may be right about that last part," so I marked him down as intelligent.
(*) Other mischievous questions to ask self-proclaimed Selfishes: "Would you sacrifice your own life to save the entire human species?" (If they notice that their own life is strictly included within the human species, you can specify that they can choose between dying immediately to save the Earth, or living in comfort for one more year and then dying along with Earth.) Or, taking into account that scope insensitivity leads many people to be more concerned over one life than the Earth, "If you had to choose one event or the other, would you rather that you stubbed your toe, or that the stranger standing near the wall there gets horribly tortured for fifty years?" (If they say that they'd be emotionally disturbed by knowing, specify that they won't know about the torture.) "Would you steal a thousand dollars from Bill Gates if you could be guaranteed that neither he nor anyone else would ever find out about it?" (Selfish libertarians only.)