Welcome!
There are still difficulties with my current model, but they're smaller than the issues I have with my best atheist theory.
What issues does your best atheist theory have?
There are a lot of bad arguments out there. Fortunately, there are also plenty of people who stand up against these arguments, which is good.
However, there is a pattern I observe quite often in such counter-arguments, which, while strictly logically valid, can become problematic later. It involves fixing all of one's counter-arguments on countering one, and only one, of the original arguer's points. I suspect that this tendency can, at best, weaken one's argument, and, at worst, allow oneself to believe things one has no intention of believing.
Let's assume, without much loss of generality, that the Wrong Argument can be expressed in the following form:
A: Some statement.
B: Some other statement.
A & B -> C: A logical inference, which, from the way B is constructed, is a fairly obvious tautology.
C: The conclusion.
Unfortunately, most of the arguments I could choose for this discussion are either highly trivial or highly controversial. I'll choose one that I hope won't cause too much trouble. Bear in mind that this is the Wrong Argument, the thing that the counter-arguer, the person presenting the good, rational refutation, is trying to demonstrate to be false. Let's designate this rational arguer as RA. The person presenting the Wrong Argument will be designated WA (Wrong Arguer).
WA: "Men have better technical abilities than women, so they should get paid more for the same engineering jobs."
WA relates a terrible sentiment, yet a pervasive one. I don't know anyone who actually espouses it in my workplace, but it was certainly commonplace not so long ago (musical evidence). Let's hope that RA has something persuasive to say against it.
Based on what I've seen of gender discussions on other forums, here's the most likely response I'd expect from RA:
RA: Don't be ridiculous! Men and women are just as well suited to technical careers as each other!
... and that's usually as far as it goes. Now, RA is right, as far as anyone knows (IANAPsychologist, though).
However, WA's argument can be broken down into the following steps:
A: Men, on average, have better technical skills than women.
B: If members of one group, on average, are better at a task than members of another group, then members of that first group should be paid more than members of the second group for performing the same work.
C: Men should be paid more than women for the same work in technical fields such as engineering.
Trivially, A & B -> C. Thus RA only needs to disprove A or B in order to break the argument. (Yes, ~A doesn't imply ~C, but WA will have a hard time proving C without A.) Both A and B are unpleasant statements that decent, rational people should probably disagree with, and C is definitely problematic.
So RA sets about attacking A. He starts by simply stating that men and women have equal potential for technical talent, on average. If WA doesn't believe that, then RA presents anecdotal evidence, then starts digging up psychological studies. Every rational discourse weapon at RA's disposal may be deployed to show that A is false. Maybe WA will be convinced, maybe he won't.
But what about B? RA has ignored B entirely in his attack on A. Now, from a strictly logical point of view, RA doesn't need to do anything with B - if he disproves A, then he disproves A & B. Attacking A doesn't mean that he accepts B as true...
... except that it kind of does.
What if WA manages to win the argument over A, by whatever means? What if WA turns out to be an evolutionary psychology clever arguer, with several papers worth of "evidence" that "proves" that men have better technical skills than women? RA might simply not have the skills or resources to refute WA's points, leading to the following exchange:
WA: Men are better engineers than women, and should be paid more!
RA: That's ridiculous. Men and women have identical potentials for technical skill!
WA: No they don't! Here are ten volumes' worth of papers proving me right!
RA: Well, gee, who am I to argue with psychology journals? I guess you're right.
WA: Glad we agree. I'll go talk to the CTO about Wanda's pay cut, shall I?
RA: Hang on a minute! Even if men are better engineers than women, that's no reason for pay inequity! Equal work for equal pay is the only fair way. If men really are better, they'll get raises and promotions on their own merit, not merely by virtue of being male.
WA: What? I spent hours getting those references together, and now you've moved the goalposts on me! I thought you weren't meant to do that!
RA: But... it's true...
WA: I think you've just taken your conclusion, "Men and women should get equal pay for the same work", and figured out a line of reasoning that gets you there. What are you, some kind of clever arguer for female engineers? Wait, isn't your mother an engineer too?
Nobody wants to be in this situation. RA really has moved the goalposts on WA, which is one of those Dark Arts that we're not supposed to employ, even unintentionally.
The problem goes deeper than simply violating good debating etiquette, though. If this debate is happening in public, then onlookers might get the impression that RA supports B. It will then be more difficult for RA to argue against B in later arguments, especially ones of the form D & B, where D is actually true. (For example, D might be "Old engineers have better technical skills than younger engineers", which is true-ish because of the benefits of long experience in an industry, but it still shouldn't mean that old engineers automatically deserve higher pay for the same work.)
Furthermore, and again IANAP, but it seems possible to me that if RA keeps arguing against A and ignoring B, he might actually start believing B. Alternatively, he might not specifically believe B, but he might stop thinking about B at all, and start ignoring the B step in his own reasoning and other people's.
So, the way to avoid all of this, is to raise all of your objections simultaneously, thusly:
WA: Men are better engineers than women, and should be paid more!
RA: Woah. Okay, first? There's no evidence to suggest that that's actually true. But secondly, even pretending for the moment that it were true, that would be no excuse for paying women less for the same work.
WA: Oh. Um. I'm pretty confident about that first point, but I never actually thought I'd have to defend the other bit. I'll go away now.
That's a best-case scenario, but it does avoid the problems above.
This post has already turned out longer than I intended, so I'll end it here. The last point I wanted to raise, though, is that an awful lot of Wrong Arguments (or good arguments, for that matter) take a form where A is an assertion of fact ("men are better engineers than women"), and B is an expression of morality ("... and therefore they should get paid more"). There are some important implications to this, for which I have a number of examples to present if people are interested.
To summarise: If someone says "A and B are true!", don't just say "A isn't true!". Say "A isn't true, and even if it were, B isn't true either!". Otherwise people might think you believe B, and they might even be right.
Welcome!
There are still difficulties with my current model, but they're smaller than the issues I have with my best atheist theory.
What issues does your best atheist theory have?
What issues does your best atheist theory have?
My biggest problem right now is all the stuff about zombies, and how that implies that, in the absence of some kind of soul, a computer program or other entity that is capable of the same reasoning processes as a person, is morally equivalent to a person. I agree with every step of the logic (I think, it's been a while since I last read the sequence), but I end up applying it in the other direction. I don't think a computer program can have any moral value, therefore, without the presence of a soul, people also have no moral value. Therefore I either accept a lack of moral value to humanity (both distasteful and unlikely), or accept the presence of something, let's call it a soul, that makes people worthwhile (also unlikely). I'm leaning towards the latter, both as the less unlikely, and the one that produces the most harmonious behaviour from me.
It's a work in progress. I've been considering the possibility that there is exactly one soul in the universe (since there's no reason to consider souls to propagate along the time axis of spacetime in any classical sense), but that's a low-probability hypothesis for now.
Welcome!
I'm planning out an actual post at the moment
Exciting! What's it about?
It's about how, if you're attacking somebody's argument, you should attack all of the bad points of it simultaneously, so that it doesn't look like you're attacking one and implicitly accepting the others. With any luck, it'll be up tonight.
Hi, I've been lurking on Less Wrong for a few months now, making a few comments here and there, but never got around to introducing myself. Since I'm planning out an actual post at the moment, I figured I should tell people where I'm coming from.
I'm a male 30-year-old optical engineer in Sydney, Australia. I grew up in a very scientific family and have pretty much always assumed I had a scientific career ahead of me, and after a couple of false starts, it's happened and I couldn't ask for a better job.
Like many people, I came to Less Wrong from TVTropes via Methods of Rationality. Since I started reading, I've found that it's been quite helpful in organising my own thoughts and casting aside unuseful arguments, and examining aspects of my life and beliefs that don't stand up under scrutiny.
In particular, I've found that reading Less Wrong has allowed, nay forced, me to examine the logical consistency of everything I say, write, hear and read, which allows me to be a lot more efficient in dicussions, both by policing my own speech and being more usefully critical of others' points (rather than making arguments that don't go anywhere).
While I was raised in a substantively atheist household, my current beliefs are theist. The precise nature of these beliefs has shifted somewhat since I started reading Less Wrong, as I've discarded the parts that are inconsistent or even less likely than the others. There are still difficulties with my current model, but they're smaller than the issues I have with my best atheist theory.
I've also had a surprising amount of success in introducing the logical and rationalist concepts from Less Wrong to one of my girlfriends, which is all the more impressive considering her dyscalculia. I'm really pleased that that this site has given me the tools to do that. It's really easy now to short-circuit what might otherwise become an argument by showing that it's merely a dispute about definitions. It's this sort of success that has kept me reading the site these past months, and I hope I can contribute to that success for other people.
Assuming rational agents with a reasonable level of altruism (by which I mean, incorporating the needs of other people and future generations into their own utility functions, to a similar degree to what we consider "decent people" to do today)...
If such a person figures that getting rid of the Nazis or the Daleks or whoever the threat of the day is, is worth a tiny risk of bringing about the end of the world, and their reasoning is completely rational and valid and altrustic (I won't say "unselfish" for reasons discussed elsewhere in this thread) and far-sighted (not discounting future generations too much)...
... then they're right, aren't they?
If the guys behind the Trinity test weighed the negative utility of the Axis taking over the world, presumably with the end result of boots stamping on human faces forever, and determined that the 3/1,000,000 chance of ending all human life was worth preventing this future from coming to pass, then couldn't Queen Victoria perform the same calculations, and conclude "Good heavens. Nazis, you say? Spreading their horrible fascism in my empire? Never! I do hope those plucky Americans manage to build their bomb in time. Tiny chance of destroying the world? Better they take that risk than let fascism rule the world, I say!"
If the utility calculations performed regarding the Trinity test were rational, altrustic and reasonably far-sighted, then they would have been equally valid if performed at any other time in history. If we apply a future discounting factor of e^-kt, then that factor would apply equally to all elements in the utility calculation. If the net utility of the test were positive in 1945, then it should have been positive at all points in history before then. If President Truman (rationally, altrustically, far-sightedly) approved of the test, then so should Queen Victoria, Julius Caesar and Hammurabi have, given sufficient information. Either the utility calculations for the test were right, or they weren't.
If they were right, then the problem stops being "Oh no, future generations are going to destroy the world even if they're sensible and altruistic!", and starts being "Oh no, a horrible regime might take over the world! Let's hope someone creates a superweapon to stop them, and damn the risk!"
If they were wrong, then the assumption that the ones performing the calculation were rational, altrustic and far-sighted is wrong. Taking these one by one:
1) The world might be destroyed by someone making an irrational decision. No surprises there. All we can do is strive to raise the general level of rationality in the world, at least among people with the power to destroy the world.
2) The world might be destroyed by someone with only his own interests at heart. So basically we might get stuck with Dr Evil. We can't do a lot about that either.
3) The world might be destroyed by someone acting rationally and altrustically for his own generation, but who discounts future generations too much (i.e. his value of k in the discounting factor is much larger than ours). This seems to be the crux of the problem. What is the "proper" value of k? It should probably depend on how much longer humans are going to be around, for reasons unrelated to the question at hand. If the world really is going to end in 2012, then every dollar spent on preventing global warming should have been spent on alleviating short-term suffering all over the world, and the proper value for k is very large. If we really are going to be here for millions of years, then we should be exceptionally careful with every resource (both material and negentropy-based) we consume, and k should be very small. Without this knowledge, of course, it's very difficult to determine what k should be.
That may be the way to avoid a well-meaning scientist wiping out all human life - find out how much longer we have as a species, and then campaign that everyone should live their lives accordingly. Then, the only existential risks that would be implemented are the ones that are actually, seriously, truly, incontrovertibly, provably worth it.
Good point. Better not draw a card if you have negative utility.
Just trust that Omega can double your utility, for the sake of argument. If you stop before you die, you get all those doublings of utility for the rest of your life.
I'd certainly draw one card. But would I stop drawing cards?
Thinking about this in commonsense terms is misleading, because we can't imagine the difference between 8x utility and 16x utility. But we have a mathematical theory about rationality. Just apply that, and you find the results seem unsatisfactory.
Thinking about this in commonsense terms is misleading, because we can't imagine the difference between 8x utility and 16x utility
I can't even imagine doubling my utility once, if we're only talking about selfish preferences. If I understand vNM utility correctly, then a doubling of my personal utility is a situation which I'd be willing to accept a 50% chance of death in order to achieve (assuming that my utility is scaled so that U(dead) = 0, and without setting a constant level, we can't talk about doubling utility). Given my life at the moment (apartment with mortgage, two chronically ill girlfriends, decent job with unpleasantly long commute, moderate physical and mental health), and thinking about the best possible life I could have (volcano lair, catgirls), I wouldn't be willing to take that bet. Intuition has already failed me on this one. If Omega can really deliver on his promise, then either he's offering a lifestyle literally beyond my wildest dreams, or he's letting me include my preferences for other people in my utility function, in which case I'll probably have cured cancer by the tenth draw or so, and I'll run into the same breakdown of intuition after about seventy draws, by which time everyone else in the world should have their own volcano lairs and catgirls.
With the problem as stated, any finite number of draws is the rational choice, because the proposed utility of N draws outweighs the risk of death, no matter how high N is. The probability of death is always less than 1 for a finite number of draws. I don't think that considering the limit as N approaches infinity is valid, because every time you have to decide whether or not to draw a card, you've only drawn a finite number of cards so far. Certainty of death also occurs in the same limit as infinite utility, and infinite utility has its own problems, as discussed elsewhere in this thread. It might also leave you open to Pascal's Scam - give me $5 and I'll give you infinite utility!
But we have a mathematical theory about rationality. Just apply that, and you find the results seem unsatisfactory.
I agree - to keep drawing until you draw a skull seems wrong. However, to say that something "seems unsatisfactory" is a statement of intuition, not mathematics. Our intuition can't weigh the value of exponentially increasing utility against the cost of an exponentionally diminishing chance of survival, so it's no wonder that the mathematically derived answer doesn't sit well with intuition.
Here's a possible problem with my analysis:
Suppose Omega or one of its ilk says to you, "Here's a game we can play. I have an infinitely large deck of cards here. Half of them have a star on them, and one-tenth of them have a skull on them. Every time you draw a card with a star, I'll double your utility for the rest of your life. If you draw a card with a skull, I'll kill you."
How many cards do you draw?
I'm pretty sure that someone who believes in many worlds will keep drawing cards until they die. But even if you don't believe in many worlds, I think you do the same thing, unless you are not maximizing expected utility. (Unless chance is quantized so that there is a minimum possible possibility. I don't think that would help much anyway.)
So this whole post may boil down to "maximizing expected utility" not actually being the right thing to do. Also see my earlier, equally unpopular post on why expectation maximization implies average utilitarianism. If you agree that average utilitarianism seems wrong, that's another piece of evidence that maximizing expected utility is wrong.
"Every time you draw a card with a star, I'll double your utility for the rest of your life. If you draw a card with a skull, I'll kill you."
Sorry if this question has already been answered (I've read the comments but probably didn't catch all of it), but...
I have a problem with "double your utility for the rest of your life". Are we talking about utilons per second? Or do you mean "double the utility of your life", or just "double your utility"? How does dying a couple of minutes later affect your utility? Do you get the entire (now doubled) utility for those few minutes? Do you get pro rata utility for those few minutes divided by your expected lifespan?
Related to this is the question of the utility penalty of dying. If your utility function includes benefits for other people, then your best bet is to draw cards until you die, because the benefits to the rest of the universe will massively outweigh the inevitability of your death.
If, on the other hand, death sets your utility to zero (presumably because your utility function is strictly only a function of your own experiences), then... yeah. If Omega really can double your utility every time you win, then I guess you keep drawing until you die. It's an absurd (but mathematically plausible) situation, so the absurd (but mathematically plausible) answer is correct. I guess.
It's instead that bad consequences result from our seeking pleasure (even if all our means-end calculations are perfectly accurate).
That sounds like a contradiction. If you're perfect at doing means-end calculations, and the best way to attain pleasure or happiness is something other than seeking it directly, then your calculations will tell you that, and you will do it.
Maybe I'm missing something, but this sounds more like an aesop about the perils of hedonism, and I'm not sure it would apply to perfect decision-makers.
Perfect decision-makers, with perfect information, should always be able to take the optimal outcome in any situation. Likewise, perfect decision-makers with limited information should always be able to choose the outcome with the best expected payoff under strict Bayesian reasoning.
However, when the actor's decision-making process becomes part of the situation under consideration, as happens when Katemega scrutinises Joe's potential for leaving her in the future, then the perfect decision-maker is only able to choose the optimal outcome if he is also capable of perfect self-modification. Without that ability, he's vulnerable to his own choices and preferences changing in the future, which he can't control right now.
I'd also like to draw a distinction between a practical pre-commitment (of the form "leaving this marriage will cause me -X utilons due to financial penalty or cognitive dissonance for breaking my vows"), and an actual self-modification to a mind state where "I promised I would never leave Kate, but I'm going to do it anyway now" is not actually an option. I don't think humans are capable of the latter. An AI might be, I don't know.
Also, what about decisions Joe made in the past (for example, deciding when he was eighteen that there was no way he was ever going to get married, because being single was too much fun)? If you want your present state to influence your future state strongly, you have to accept the influence of your past state on your present state just as strongly, and you can't just say "Oh, but I'm older and wiser now" in one instance but not the other.
Without the ability to self-modify into a truly sincere state wherein he'll never leave Kate no matter what, Joe can't be completely sincere, and (by the assumptions of the problem) Kate will sense this and his chances of his proposal being accepted will diminish. And there's nothing he can do about that.
It's an interesting situation, and I can see the parallel to Newcombe's Problem. I'm not certain that it's possible for a person to self-modify to the extent that he will never leave his wife, ever, regardless of the very real (if small) doubts he has about the relationship right now. I don't think I could ever simultaneously sustain the thoughts "There's about a 10% chance that my marriage to my wife will make me very unhappy" and "I will never leave her no matter what". I could make the commitment financially - that, even if the marriage turns awful, I will still provide the same financial support to her - but not emotionally. If Joe can modify his own code so that he can do that, that's very good of him, but I don't think many people could do it, not without pre-commitment in the form of a marital contract with large penalties for divorce, or at least a very strong mentality that once the vows are said, there's no going back.
Perhaps the problem would be both more realistic and more mathematically tractable if "sincerity" were rated between 0 and 1, rather than being a simple on/off state? If 1 is "till death do us part" and 0 is "until I get a better offer", then 0.9 could be "I won't leave you no matter how bad your cooking gets, but if you ever try to stab me, I'm out of here". Then Kate's probability of accepting the proposal could be a function of sincerity, which seems a much more reasonable position for her.
Could this be an example where rationality and self-awareness really do work against an actor? If Joe were less self-aware, he could propose with complete sincerity, having not thought through the 10% chance that he'll be unhappy. If he does become unhappy, he'd then feel justified in this totally unexpected change inducing him to leave. The thing impeding Joe's ability to propose with full sincerity is his awareness of the possibility of future unhappiness.
Also, it's worth pointing out that, by the formulation of the original problem, Kate expects Joe to stay with her even if she is causing him -125 megautilons of unhappiness by forcing him to stay. That seems just a touch selfish. This is something they should talk about.
View more: Next
It's hard to build intuitions about the moral value of intelligent programs right now, because there aren't any around to talk to. But consider a hypothetical that's as close to human as possible: uploads. Suppose someone you knew decided to undergo a procedure where his brain would be scanned and destroyed, and then a program based on that scan was installed on a humanoid robot body, so that it would act and think like he did; and when you talked to the robot, he told you that he still felt like the same person. Would that robot and the software on it have moral value?
I agree, intuition is very difficult here. In this specific scenario, I'd lean towards saying yes - it's the same person with a physically different body and brain, so I'd like to think that there is some continuity of the "person" in that situation. My brain isn't made of the "same atoms" it was when I was born, after all. So I'd say yes. In fact, in practice, I would definitely assume said robot and software to have moral value, even if I wasn't 100% sure.
However, if the original brain and body weren't destroyed, and we now had two apparently identical individuals claiming to be people worthy of moral respect, then I'd be more dubious. I'd be extremely dubious of creating twenty robots running identical software (which seems entirely possible with the technology we're supposing) and assigning them the moral status of twenty people. "People", of the sort deserving of rights and dignity and so forth, shouldn't be the sort of thing that can be arbitrarily created through a mechanical process. (And yes, human reproduction and growth is a mechanical process, so there's a problem there too.)
Actually, come to think of it... if you have two copies of software (either electronic or neuron-based) running on two separate machines, but it's the same software, could they be considered the same person? After all, they'll make all the same decisions given similar stimuli, and thus are using the same decision process.