Once upon a time, I met someone who proclaimed himself to be purely selfish, and told me that I should be purely selfish as well.  I was feeling mischievous(*) that day, so I said, "I've observed that with most religious people, at least the ones I meet, it doesn't matter much what their religion says, because whatever they want to do, they can find a religious reason for it.  Their religion says they should stone unbelievers, but they want to be nice to people, so they find a religious justification for that instead.  It looks to me like when people espouse a philosophy of selfishness, it has no effect on their behavior, because whenever they want to be nice to people, they can rationalize it in selfish terms."

And the one said, "I don't think that's true."

I said, "If you're genuinely selfish, then why do you want me to be selfish too?  Doesn't that make you concerned for my welfare?  Shouldn't you be trying to persuade me to be more altruistic, so you can exploit me?"

The one replied:  "Well, if you become selfish, then you'll realize that it's in your rational self-interest to play a productive role in the economy, instead of, for example, passing laws that infringe on my private property."

And I said, "But I'm a small-L libertarian already, so I'm not going to support those laws.  And since I conceive of myself as an altruist, I've taken a job that I expect to benefit a lot of people, including you, instead of a job that pays more.  Would you really benefit more from me if I became selfish?  Besides, is trying to persuade me to be selfish the most selfish thing you could be doing?  Aren't there other things you could do with your time that would bring much more direct benefits?  But what I really want to know is this:  Did you start out by thinking that you wanted to be selfish, and then decide this was the most selfish thing you could possibly do?  Or did you start out by wanting to convert others to selfishness, then look for ways to rationalize that as self-benefiting?"

And the one said, "You may be right about that last part," so I marked him down as intelligent.

(*)  Other mischievous questions to ask self-proclaimed Selfishes:   "Would you sacrifice your own life to save the entire human species?"  (If they notice that their own life is strictly included within the human species, you can specify that they can choose between dying immediately to save the Earth, or living in comfort for one more year and then dying along with Earth.)  Or, taking into account that scope insensitivity leads many people to be more concerned over one life than the Earth, "If you had to choose one event or the other, would you rather that you stubbed your toe, or that the stranger standing near the wall there gets horribly tortured for fifty years?"  (If they say that they'd be emotionally disturbed by knowing, specify that they won't know about the torture.)  "Would you steal a thousand dollars from Bill Gates if you could be guaranteed that neither he nor anyone else would ever find out about it?"  (Selfish libertarians only.)

New Comment
72 comments, sorted by Click to highlight new comments since:
Some comments are truncated due to high volume. (⌘F to expand all)Change truncation settings

Have you read Mark Twain's "What is Man"? If I recall correctly, there he lays out his argument that man is already always selfish. For example, we do good deeds ultimately for our own comfort, because they make us feel good. (If I recall rightly, he also makes the, to me rather more interesting, point that people are born happy or sad rather than are made happy or sad by specific, supposedly uplifting or depressing, thoughts. That seems to anticipate the modern pharmaceutical approach to mood regulation.)

For my part, I think that "selfish&q... (read more)

Did Hopefully Anonymous figure this out and stop expending effort commenting or posting on his anonymous blog?

So, it seems that Eliezer's working definition of an intelligent person is "someone who agrees with me".

I must point out that "whenever they want to be nice to someone" entails a desire to be nice to someone. Your very phrase defines it as being in their interests to be nice to someone. Rationalization isn't even necessary here. You wanted to do something - you did it. Selfishness isn't that complicated.

My guess would be that this individual had read Atlas Shrugged and hadn't fully understood what selfish meant in the context. Ayn Rand was setting out to redefine the word, not to glorify the "old" meaning.

[-]Tiiba2110

I think that people are born selfish. This is based on the fact that, if I grew up in an environment that didn't compel me to be either selfish or selfless, I'd probably be selfish. Babies are selfish. I value those creatures that I was convinced to value. Slave owners taught their children that it's okay to beat slaves, and the children were happy to comply. Now most people disregard the pain of food animals because they can get away with it.

Of course, some of my actions are genuinely altruistic. I chose to give up meat, although this brings me little tangible benefit. (It does get me out of some accusations of hypocricy.) One reason why I let myself become like this is that in human society, being nice is a habit that keeps my ass from getting kicked. And it needs to be a habit, because I'm not smart enough to delude everybody that I care, when I actually see them all as obstacles.

If I somehow become so powerful that I no longer depend on anyone, and noting they do can harm me, I will probably quickly become corrupted by my power.

But I agree that now, I can't be considered purely selfish.

"So, it seems that Eliezer's working definition of an intelligent person is "someone who agrees with me"."

My definition of an intelligent person is slowly becoming "someone who agrees with Eliezer", so that's all right. Plus, the guy showed ability to revise a strongly held belief.

Read the comments at Hopefully Anonymous' most recent post. He explains why he has been inactive.

I want you to be altruistic, Eliezer. That's partly because I think you're intelligent. I would prefer if some people were more selfish though.

I choose living in comfort for one more year. There are things I might die for, but I don't know what exactly. Perhaps to spite someone. If other people knew that I had the chance to save the world and were going to punish me for failing to do so, I might not risk their wrath. I also choose the stanger getting tortured, ... (read more)

"My definition of an intelligent person is slowly becoming 'someone who agrees with Eliezer', so that's all right."

That's not in the spirit of this blog. Status is the enemy, only facts are important.

[-]Tiiba2150

"That's not in the spirit of this blog. Status is the enemy, only facts are important."

See? Another smart man agrees with Eliezer. That's what I'm talking about.

Almost as though Eliezer isn't a person, but a system of thought.

Taking a cue from some earlier writing by Eli, I suppose one way to give ethical systems a functional test is to imagine having access to a genie. An altruist might ask the genie to maximize the amount of happiness in the universe or something like that, in which case the genie might create a huge number of wireheads. This seems to me like a bad outcome, and would likely be seen as a bad outcome by the altruist who made the request of the genie. A selfish person might say to the genie "create the scenario I most want/approve of." Then it would be... (read more)

3NaomiLong
It seems like this is based more on the person's ability to optimize. The altruistic person who realized this flaw would then be able to (assuming s/he had the intelligence and rationality to do so) calculate the best possible wish to benefit the most number of people.
-1ragnarrahl
Notice how you had to assume the altruist to have the extraordinary degree of intelligence and rationality to calculate the best possible wish and Stephen merely had to assume that the selfishness was of the goodwill-toward-men-if-it-doesn't-cost-me-anything sort? When you require less implausible assumptions to render a given ethical philosophy genie-resilient, the philosophy is more genie-resilient.

"Do not seek to follow in the footsteps of the wise, seek what they sought." -- Nanzan Daishi, quoted by Matsuo Basho.

You most certainly are right. He is a fool who disagrees.

Stephen: the altruist can ask the Genie the same thing as the selfish person. In some sense, though, I think these sorts of wishes are 'cheating,' because you are shifting the computational/formalization burden from the wisher to the wishee. (Sorry for the thread derail.)

"An altruist might ask the genie to maximize the amount of happiness in the universe or something like that, in which case the genie might create a huge number of wireheads. This seems to me like a bad outcome, and would likely be seen as a bad outcome by the altruist who made the request of the genie."

Eh? An altruist would voluntarily summon disaster upon the world?

By the way, I have some questions about wireheading. What is it, really? Why is it so repulsive? Is it really so bad? If, when you imagine your brain rewired, you envision something t... (read more)

0ragnarrahl
" Eh? An altruist would voluntarily summon disaster upon the world?" No, an altruist's good-outcomes are complex enough to be difficult to distinguish from disasters by verbal rules. An altruist has to calculate for 6 billion evaluative agents, an egoist just 1. " By the way, I have some questions about wireheading. What is it, really? Why is it so repulsive?" Wireheading is more or less where a sufficiently powerful agent told to optimize for happiness optimizes for the emotional referents without the intellectual and teleological human content typically associated with that. You can perform primitive wireheading right now with various recreational drugs. The fact that almost everyone uses at least a few of the minor ones tells us that wireheading isn't in and of itself absolutely repugnant to everyone-- but the fact that only the desperate pursue the more major forms of wireheading available and the results (junkies) are widely looked upon as having entered a failure-mode is good evidence that it's not a path we want sufficiently powerful agents to go down. " it's possible that an AI ordered to make you happy would choose some other course of action. " When unleashing forces one cannot un-unleash, one wants to deal in probability, not possibility. That's more or less the whole Yudkowskian project in a nutshell.

Humans seem to gain social status by persuading others to agree with them. This is one of the reasons we resist being persuaded by good arguments. So a selfish person who wanted social status could want you to be selfish also in order to gain social status, showing our dominance via the submission of others.

2EniScien
But won't a real egoist try to convince you for self-affirmation in a position that is beneficial to him and disadvantageous to you, and not of egoism?

It seems to me that this discussion is somewhat misleading. Each one of us members of the Homo Sapiens species operates to pursue his/her own interests, as dictated by her/his genetic code. But in order to do so, we have got to cooperate with some of our fellow human beings. Each and every one of our actions is in some ways a combination of these two typical aspects of our behavior. There is no such thing as a totally selfish or a totally unselfish behavior/action/activity, much less can we talk of a totally selfish or a totally unselfish person -- only ex... (read more)

In my above comment at 5:49 AM, the sentence "Whether we chose to declare ourselves selfish or unselfishness" should actually read: "Whether we chose to declare ourselves selfish or unselfish"

I said, "If you're genuinely selfish, then why do you want me to be selfish too? Doesn't that make you concerned for my welfare? Shouldn't you be trying to persuade me to be more altruistic, so you can exploit me?"

The objection you press against your interlocutor was anticipated by Max Stirner, the renowned champion of egoism, who replied as follows:

Do I write out of love to men? No, I write because I want to procure for my thoughts an existence in the world; and, even if I foresaw that these thoughts would deprive you of your rest and your pea
... (read more)

Selfishness - or to avoid confusion, let's call the concept 'self-interest' - takes on rather a different appearance when it's realized that the 'self' is not something necessarily limited to the boundaries of the physical form that embodies the distinction.

To the degree that we identify with and value the rest of humanity, sacrificing one's own existence to preserve the rest of humanity can be in the self-interest. To the degree that we don't, or that we negatively value the rest of humanity, that action can be against self-interest. If we disliked humanity enough, we'd choose to destroy it even if it cost us our own lives (which presumably we'd value) in the process.

So, it seems that Eliezer's working definition of an intelligent person is "someone who agrees with me".

Communities form because they repress the incompatible. Particularly on the Internet, where it's easy to restrict who can participate, people tend to agree not because they persuade one another but because they seek out and associate with like-minded people.

Obviously Eliezer thinks that the people who agree with the arguments that convince him are intelligent. Valuing people who can show your cherished arguments to be wrong is very nearly a post-human trait - it is extraordinarily rare among humans, and even then unevenly manifested.

[-]ehj200

If we posit an ideal world where every person has perfect and complete knowledge, and the discipline and self-control to act consistently on that knowledge, it's possible we can equate the most self-interested act with the most ethical.

Until we have that ideal world, to posit that people always simply do what they want to do anyway, and rationalize their behavior to their philosophy of life, is to engage in a bit of the same rationalization when we conclude that "if only they knew as much as I know, they would do what I think they should do."

Too ... (read more)

Are providing answers to questions like "Would you do incredible thing X if condition Y was true" really necessary if thing X is something neither person would likely ever be able to do and condition Y is simply never going to happen? It seems easy to construct impossible moral challenges to oppose a particular belief, but why should beliefs be built around impossible moral edge cases? Shouldn't a person be able to develop a rational set of beliefs that do fail under extreme moral cases, but at the same time still hold a perfectly strong and not contradictory position?

Obviously Eliezer thinks that the people who agree with the arguments that convince him are intelligent. Valuing people who can show your cherished arguments to be wrong is very nearly a post-human trait - it is extraordinarily rare among humans, and even then unevenly manifested.

On the other hand, if we are truly dedicated to overcoming bias, then we should value such people even more highly than those whom we can convince to question or abandon their cherished (but wrong) arguments/beliefs.

The problem is figuring out who those people are.

But it's very di... (read more)

It's not that the one agreed with me and declared himself no longer selfish, but that he showed nonzero reactivity in the face of an unexpected argument, a rare thing. Further conversation (not shown) did seem to show that he was thinking about it. You don't see, say, Caledonian ever updating his views, or showing nonzero dependency between what I say and his ability to comment negatively on every post.

Ah, but I have yet to be confronted with an argument that would cause me to update my views.

So now the problem expands: if two people disagree about the worth of an argument, what criteria do we use to choose between them? What if we ARE one of the two people? How do we use our own judgement to evaluate our own judgement?

Simply, we can't. At best, we can look for simple errors that we presume we can objectively determine, and leave the subtle issues to other systems.

I don't see you admitting you were wrong in the previous threads, Eliezer. Should I interpret that as your unwillingness to admit to error, or that you're so much smarter than me that I can't even comprehend how you're actually correct?

"Selfish" in the negative sense means not just pursuing one's own interests, but doing so heedless of the harm one's actions may be causing others. I don't think there are many proponents of "selfishness" in this sense.

There are people that are "selfless" in the sense that they not only don't act according to their direct self-interest, they even abandon their own concepts of true and false, right and wrong, trusting some external authority to make these judgments for them. Religious, political, whatever. People who praise selfishness are generally contrasting it with this kind of selflessness.

My understand is that the philosophy of rational self-interest, as forwarded by the Objectivists, contains a moral system founded first on the pursuit of maintaining a high degree of "conceptual" volitional consciousness and freedom as a human being. Anything that robs one's life or robs one's essential humanity is opposed to that value. The Objectivist favor of capitalism stems from a belief that capitalism is a system that does much to preserve this value (the essential freedom and humanity of individuals). Objectivists are classical libertaria... (read more)

[-]GNZ00

like pablo I have considered whether arguing for in my case altruism/utilitarianism is always altruistic and thought "well probably" - but I dont analyse it much because in the end I don't know if it would matter - it seems I do what I do because 'that is what I am', more than 'that is what I think is right'. I guess it works the other way too eh.

Michael Sullivan:

That's an exceptionally clear exegesis. Thanks!

Pablo Stafforini:

The words of Max Stirner (with whom I am admittedly unfamiliar) that you quote seem to me like so much bluster and semantic question-begging.

Do I write out of love to men? No, I write because I want to procure for my thoughts an existence in the world; and, even if I foresaw that these thoughts would deprive you of your rest and your peace, even if I saw the bloodiest wars blah blah blah

He sings not out of love for the hearer, but because he loves to sing and the hearer is us... (read more)

2buybuydandavis
I'm a fan of Stirner, but have always found this particular passage disingenuous. Certainly he sings because he is a singer. It's possible that he is unconcerned with the effect of his song on those without the ears to hear it, but I don't find it credible that he had no hope of benefiting those with the ears to hear. So I think he writes, at least somewhat, out of love for at least some men. Because the usual forms of unselfishness are based in conceptual confusions that make you less useful to me, and often downright dangerous. And, it's both sad and rather distasteful to watch you live your life in such a crippled fashion. Such waste offends my sensibilities. I would much prefer that my neighbors live for themselves, than live for God, Gaia, Allah, Evolution, Justice, the State, The Volk, The Proletariat, etc.
0TheOtherDave
I can see where this might be true, but I can also see where it might be mere sophistry concealing a fundamental concern for your neighbor's well-being. Can you provide some concrete examples of typical ways in which your neighbors' unselfishness lessens their usefulness to you?
3buybuydandavis
I guess I didn't make myself clear. I'm not concealing a fundamental concern for my neighbor's well being. I have it, and I think Stirner does too, despite his disingenuous denial here. Hence my comment that he writes, at least in part, out of love for some men. Stirner, The Ego and It's Own: It's the commandment of love he rejects, not his own love. Loving other men is no more unselfish than loving your car. You like it shiny and running well, maybe even after you sell it to someone else. It is perfectly selfish to love what you love, and hate what you hate. To care about what you care about. Why should I limit my concerns to what lies in a 1 inch bubble around myself? It's not what concerns you that makes you an egoist, it's whether you bow to an ideological compulsion to serve an alien concern over your own. My own take on Stirner's Egoism is that it is best distinguished as the antidote to various forms of Moral Objectivism, not Altruism.
1taelor
Note that Stirner believed that it is impossible to serve an alien concern over your own. That fact that you are concerned with something makes it your concern. Stirner called people who claim to serve an alien concern above their own "involuntary egoists", and found the entire state of affairs to be laughably absurd.
0antigonus
This makes little sense to me. Other people, unlike cars, have interests; and loving other people tends to have the effect of causing one to adopt those interests as one's own. What exactly is unselfishness supposed to look like, if not that?

Daniel, you can read Stirner's book here. It's not really "macho", that would be more Ragnar Redbeard's "Might Makes Right".

Among the things Stirner writes about is freedom of expression. He does not care for what the state or the church say he may write, because he takes such freedom for himself (many unauthorized printings of his book were made). He does not respect the holy but instead regards taboos against blasphemy as attempted restrictions on him that he will violate. For people who say that he ought not to speak of certain thing... (read more)

TGGP:

Thanks for the tip. I'll check it out.

Eliezer's descriptions of his intellectual history and past mistakes are very convincing positive signals to me.

I agree, but have a nagging doubt. When I read years-old writings where he makes some of those mistakes he sounds about as knowledgeable, just as smart, just as honest, and just as sure he's right as he is now in his new beliefs.

Although I was convinced by very few of those older mistakes (before I searched and found retractions) that could just as easily mean new arguments got super persuasive rather than super accurate.

His writings have convinc... (read more)

PS. I know arguments should be judged based on their own worth and not who made them, but there are other factors.

Sometimes people change their minds because new evidence and novel arguments have forced them to re-evaluate their positions.

Other times, people have never thought through even the obvious arguments, or are convinced easily by weak data and unimpressive theses, so they shift from one point to another.

If someone changes their positions frequently, AND they're very confident in their positions, that's a bad sign.

Eliezer: “Other mischievous questions to ask self-proclaimed Selfishes: "Would you sacrifice your own life to save the entire human species?" (If they notice that their own life is strictly included within the human species, you can specify that they can choose between dying immediately to save the Earth, or living in comfort for one more year and then dying along with Earth.)”

What If the person saving the entire human species wanted to commit suicide anyway or if he/she has this dream of heroism and hopes to become immortal and be rewarded in an afterlife. An apparently selfless deed can be a very selfish one.

The question Eliezer raises is the first problem any religious person has to face once he abandons the god thesis, i.e. why should I be good now? The answer, I believe, is that you cannot act contrary to your genetic nature. Our brains are wired (or have modules in Pinker terms) for various forms of altruism, for group survival reasons probably. I therefore can’t easily commit acts against my genetic nature, even if intellectually I can see they are in my best interests. (As Eliezer has already recognised this is why AI or uploaded personalities are so dan... (read more)

If someone changes their positions frequently, AND they're very confident in their positions, that's a bad sign.

I don't think he changes his mind too frequently, and being overly confident at the time of now-abandoned positions isn't unusual. My point was that in Eliezer's case a knowledgeable, smart, honest, and self-certain argument doesn't imply strong evidence of truth, because those qualities appear in arguments that turned out false.

To be honest I think I was hoping someone would leap to his defense and crush my argument, giving me permission to be... (read more)

ChrisA, why are you intellectually convinced you should be selfish? Rationality doesn't demand any particular goals. A genuinely altruistic person, if uploaded, would overwrite the "brain code" (a bad analogy; evolved tendencies aren't deterministic "code") that promotes selfishness.

Recovering,

While I wish I had something reassuring to say on this subject, you should probably be quite disturbed if you find my work from 1997 sounding as persuasive as my work from 2007.

Nick

My response is, evolution! Let's say a genuinely (what ever that means) altruistic entity exists. He then is uploaded. He then observed that not all entities are fully altruistic, in other words they will want to take resources from others. In any contest over resources this puts the altruistic entity at a disadvantage (he is spending resources helping others that he could use to defend himself). With potentially mega intelligent entities any weakness is serious. He realises that very quickly he will be eliminated if he doesn't fix this weakness. He ei... (read more)

While I wish I had something reassuring to say on this subject, you should probably be quite disturbed if you find my work from 1997 sounding as persuasive as my work from 2007

But I said...

Although I was convinced by very few of those older mistakes (before I searched and found retractions) that could just as easily mean new arguments got super persuasive rather than super accurate.

All my comments mean in practice is that even though once I study and investigate your (new) arguments they nearly always seem to me to be right, I won't let myself start lazily... (read more)

Fair enough, Recovering. My own point is that:

When I read years-old writings where he makes some of those mistakes he sounds about as knowledgeable, just as smart, just as honest, and just as sure he's right as he is now in his new beliefs.

Then those factors aren't very good discriminators of truth, are they? It's not just "improper" to take them into account, it actually doesn't work.

In whatever facets I sounded about as "knowledgeable", "smart", "honest", or "self-assured" then as now, you might take th... (read more)

In whatever facets I sounded about as "knowledgeable", "smart", "honest", or "self-assured" then as now, you might take these facets into account in deciding whether someone's arguments are worth your time to read, but you shouldn't take them into account in deciding whether the person is right.

Agreed. Having said that, I do find those facets to correlate with truth, but the correlation flattens out for high values. Besides, the first two would be hard for me to judge well between your 1997 and 2007 selves, for obvio... (read more)

When I say that I am selfish, I mean to express that I think the best model for "altruism" is that of a good consumed much like any other. I consume it for my personal enjoyment, not in proportion to the benefit received by the recipient. And, ceteris paribus, with a declining marginal utility as quantity consumed increases.

In my eyes, a "true" altruist would value the 1000th meal provided to a starving third-worlder on par with the 1st one provided, given that the beneficiaries valued the meals similarly. Nobody behaves that way. Altru... (read more)

Mark, altruists have to deal with their costs too.

It's possible for an altruist to value the thousandth altruistic meal as much as the first, but as his resources shrink the value of the alternatives rises. If I provide meals for a hundred thousand starving people and then I have nothing left and I become a starving person myself, that isn't good. At some point I want to keep enough capital to maintain my continuing ability to feed starving people.

I'm not claiming that it's true that no altruist experiences diminishing returns, or even that there is an altruist who doesn't experience diminishing returns. But the behavior doesn't prove that there couldn't be, and so this isn't a killing blow.

J Thomas,

A non-diminishing-returns altruist would hit a point where the utility of spending a marginal resource on a "selfish" purpose dips below the best use of that resource for an "altruistic" purpose. Every single marginal resource after that should go toward altruistic purposes as well. Why? Because for anyone with non-astronomical resources, there is effectively an endless supply of altruistic options that all provide effectively the same degree of benefit to the recipient. The non-diminishing-returns altruist would increase altru... (read more)

As a selfish prizefighter, I want to beat my opponent. If I was an altruist instead, I don't think I'd be able to win one fight. Because I am in fact selfish, fighting an opponent who is an altruist would not do much for my self-esteem. Only in fighting better fighters than I am do I learn, not by fighting someone inferior. If a superior fighter does not do his best in a given match with me for some reason, I cannot objectively pretend to be better than him just because I won once. It benefits me to beat him when he's at his best. I like to share my knowle... (read more)

Re: It looks to me like when people espouse a philosophy of selfishness, it has no effect on their behavior

This is almost certainly not true of conscious genetic selfishness. Such individuals can be expected to engage in various rare and unusual activities - such as donating to sperm banks.

[-][anonymous]10

There's all kinds of incoherency with the idea of "selfishness" being defined as "acting in your own interest." What is your own interest? Can't you define it circularly as being anything you desire to do? Being a "selfish person" doesn't necessarily make sense by that definition.

Maybe it's better to go the opposite direction. Selfishness is indifference to the desires and well-being of others. Whatever else you may be doing (let's drop the question of whether it's "self-interested" or not), it's more important ... (read more)

I'm relieved that at least a few people mentioned Stirner. The selfishness EY portrays is not representative of selfishness in the sense any of the literature of Philosophical Egoism of which I am aware. His scenario doesn't even correspond to what a Randian might believe about selfishness.

Max Stirner is the best and most intellectually consistent egoist I'm aware of. Less accomplished but more polemical writers in the same vein include John L. Walker, Benjamin Tucker, John Badcock, Dora Marsden (for a period of time), and Sid Parker.

Max Stirner's "... (read more)

I've always felt that both "selfishness" and "altruism" were results of the Fundamental Attribution Error. Some actions are deemed "selfish" according to society's mean value set, others are deemed "altruistic" or "selfless". Personally, I'm more interested in the chain of events that ultimately lead up to an action being performed and in the chain of events that occurred as a result of it than I am in applying labels of dubious and limited utility to things.

[-][anonymous]00

"Would you steal a thousand dollars from Bill Gates if you could be guaranteed that neither he nor anyone else would ever find out about it?"

Without making a case for or against my own altruism, I'd definitely answer yes to this question, for two reasons, which are really one reason:

1) From a utilitarian standpoint, a thousand dollars benefits me much more than losing a thousand dollars harms Bill Gates. Much, much, much more.

2) I am destitute and in grave need of money

Is this fair reasoning? I'd also say the monetary status of the individual ... (read more)

[This comment is no longer endorsed by its author]Reply

"If you're genuinely selfish, then why do you want me to be selfish too? Doesn't that make you concerned for my welfare? Shouldn't you be trying to persuade me to be more altruistic, so you can exploit me?"

You're conflating selfishness with vulgar egoism. Suppose your well-being makes me happy, and I believe that making you selfish will make you happier. Then convincing you to be selfish is the self-interested thing to do. If I tried to convince you to be more altruistic (in the self-sacrificing sense, not in the benevolent sense) so I could... (read more)

If I didn't know that someone was going to be tortured, I would rather stub my toe, and I do not claim to be selfish. Otherwise I am not really sure how to interpret the question.

Many big-L Libertarians I've met - along with those who consider themselves to be trench-fighters for Ayn Rand-ian Objectivism - seem to want to conflate "selfishness" with "enlightened self-interest" for the positive connotations of the latter... yet their rationale for various big-L proposals (such as "let's turn over national security to corporations, who will certainly never abuse the power to force decisions upon people") tends to be of the extremely rosy, happy death spiral, declare-anything-that-doesn't-fit-an-"ext... (read more)

1Shmi
If you have a look at this blog post by one of the more famous ex-regulars, this is basically the "motte-and-bailey" tactics, where motte is "selfishness = "enlightened self-interest" and bailey is something like "let the free market rule".
[-][anonymous]-20

I think this whole egoism vs. altruism debate is too much black and white thinking.

What about evolutionary ethics? Doing what is good for your genes and the survival of the species.

Therefore, one would put self-interest over the interest of other people, but not totally disregard the latter. Also some people like your children would be more important than others. The ranking would look something like this:

  1. Yourself
  2. Your children
  3. Your family and relatives
  4. The rest of the world

Selfishness is the default, but can be overridden if the bad consequences times... (read more)

I used to think about self-proclaimed selfishness in this way all the time, but now I think I "get" the doctrine of selfishness. For instance, the answer to the first mischevous question is actually "because I get personal happiness from preaching selfishness, and that vastly overshadows any happiness I could instead get from a single person's altruism". The answer to own life vs. human species is "because a guilt-ridden life wouldn't maximise my personal happiness", and so on.

When we weigh options in our mind, we pick the one that yields the cocktail of chemicals/neurotransmitters that induce the strongest positive response in our reward center. Or rather, the cocktail of chemicals/neurotransmitters that elicits the strongest positive response Is able to pass its signals through to the motor neurons.

A desire to be moral, a desire to avoid pain, a desire to protect kin, all release chemicals.

Seen in this light, the phrase "everything one does is selfish" appears to reduce to "all choices are weighed through one's... (read more)

The one replied:  "Well, if you become selfish, then you'll realize that it's in your rational self-interest to play a productive role in the economy, instead of, for example, passing laws that infringe on my private property."

 

I get that this isn't the point of the post but... what? No, just no. If I'm selfish I'm going to pass laws that don't infringe on MY private property, why should I care about yours? Indeed, I'll just go on with using any shred of political influence I have to make sure your taxes end up paying my expenses and thus increas... (read more)

[-][anonymous]20

My answer to your first mischievous question tends to be: "if you (also) identify as selfish, this will make you more predictable and thereby more trustworthy".  

Don't give two shits about one guy's contribution to the economy. My selfish incentives are purely local.

Besides, it's always better to cooperate with a rich guy than exploit a poor one.

But of course there's no need to convince you to be something you already are

"f they say that they'd be emotionally disturbed by knowing, specify that they won't know about the torture."

Couldn't one argue that having preferences about things u assume u dont know, wouldn't affect your actions?

When I'm deciding on an actual action, I can only take into account things I know, and nothing else?

So the preference of the case where I would never know about the person being tortured couldn't affect my actions, so in that sense doesn't matter?

You can't expect a truly selfish agent to answer any of these questions honestly. Admitting you would steal in certain circumstances would cause them reputational damage. Most of us have programming in place to prevent us from incriminating ourselves morally.