All of thrawnca's Comments + Replies

A: Live 500 years and then die, with certainty. B: Live forever, with probability 0.000000001%; die within the next ten seconds, with probability 99.999999999%

If this was the only chance you ever get to determine your lifespan - then choose B.

In the real world, it would probably be a better idea to discard both options and use your natural lifespan to search for alternative paths to immortality.

3entirelyuseless
I disagree, not surprisingly, since I was the author of the comment to which you are responding. I would choose A, and I think anyone sensible would choose A. There's not much one can say here in the way of argument, but it is obvious to me that choosing B here is following your ideals off a cliff. Especially since I can add a few hundred 9s there, and by your argument you should still choose B.

somewhat confident of Omega's prediction

51% confidence would suffice.

  • Two-box expected value: 0.51 $1K + 0.49 $1.001M = $491000
  • One-box expected value: 0.51 $1M + 0.49 $0 = $510000

At some point you'll predictably approach death

I'm pretty sure that decision theories are not designed on that basis. We don't want an AI to start making different decisions based on the probability of an upcoming decommission. We don't want it to become nihilistic and stop making decisions because it predicted the heat death of the universe and decided that all paths have zero value. If death is actually tied to the decision in some way, then sure, take that into account, but otherwise, I don't think a decision theory should have "death is inevitably coming for us all" as a factor.

3Wes_W
You are wrong. In fact, this is a totally standard thing to consider, and "avoid back-chaining defection in games of fixed length" is a known problem, with various known strategies.

How do you resolve that tension?

Well, as previously stated, my view is that the scenario as stated (single-shot with no precommitment) is not the most helpful hypothetical for designing a decision theory. An iterated version would actually be more relevant, since we want to design an AI that can make more than one decision. And in the iterated version, the tension is largely resolved, because there is a clear motivation to stick with the decision: we still hope for the next coin to come down heads.

0hairyfigment
Are you actually trying to understand? At some point you'll predictably approach death, and predictably assign a vanishing probability to another offer or coin-flip coming after a certain point. Your present self should know this. Omega knows it by assumption.

what happens when the consequences grow large? Say 1 person to save 500, or 1 to save 3^^^^3?

If 3^^^^3 lives are at stake, and we assume that we are running on faulty or even hostile hardware, then it becomes all the more important not to rely on potentially-corrupted "seems like this will work".

Well, humans can build calculators. That they can't be the calculators that they create doesn't demand an unusual explanation.

Yes, but don't these articles emphasise how evolution doesn't do miracles, doesn't get everything right at once, and takes a very long time to do anything awesome? The fact that humans can do so much more than the normal evolutionary processes can marks us as a rather significant anomaly.

0hairyfigment
Not really. Birds can fly better than evolution can. As far as intelligence goes, we're far from the only animals who can make tools. Since this typically takes less than a year, they're already faster than most versions of the mindless process called evolution.

Your decision is a result of your decision theory

I get that that could work for a computer, because a computer can be bound by an overall decision theory without attempting to think about whether that decision theory still makes sense in the current situation.

I don't mind predictors in eg Newcomb's problem. Effectively, there is a backward causal arrow, because whatever you choose causes the predictor to have already acted differently. Unusual, but reasonable.

However, in this case, yes, your choice affects the predictor's earlier decision - but since th... (read more)

3Wes_W
Yes, that is the problem in question! If you want the payoff, you have to be the kind of person who will pay the counterfactual mugger, even once you no longer can benefit from doing so. Is that a reasonable feature for a decision theory to have? It's not clear that it is; it seems strange to pay out, even though the expected value of becoming that kind of person is clearly positive before you see the coin. That's what the counterfactual mugging is about. If you're asking "why care" rhetorically, and you believe the answer is "you shouldn't be that kind of person", then your decision theory prefers lower expected values, which is also pathological. How do you resolve that tension? This is, once again, literally the entire problem.

Humans can do things that evolutions probably can't do period over the expected lifetime of the universe.

This does beg the question, How, then, did an evolutionary process produce something so much more efficient than itself?

(And if we are products of evolutionary processes, then all our actions are basically facets of evolution, so isn't that sentence self-contradictory?)

0Crux
The evolutionary process produced humans, and humans can create certain things that evolution wouldn't have been able to produce without producing something like humans to indirectly produce those things. Your question is no more interesting than, "How could humans have built machines so much faster at arithmetic than themselves?" Well, humans can build calculators. That they can't be the calculators that they create doesn't demand an unusual explanation.

there is no distinction between making the decision ahead of time or not

Except that even if you make the decision, what would motivate you to stick to it once it can no longer pay up?

Your only motivation to pay is the hope of obtaining the $10000. If that hope does not exist, what reason would you have to abide by the decision that you make now?

3Wes_W
Your decision is a result of your decision theory, and your decision theory is a fact about you, not just something that happens in that moment. You can say - I'm not making the decision ahead of time, I'm waiting until after I see that Omega has flipped tails. In which case, when Omega predicts your behavior ahead of time, he predicts that you won't decide until after the coin flip, resulting in hypothetically refusing to pay given tails, so - although the coin flip hasn't happened yet and could still come up heads - your yet-unmade decision has the same effect as if you had loudly precommitted to it. You're trying to reason in temporal order, but that doesn't work in the presence of predictors.

I didn't mean to suggest that the existence of suffering is evidence that there is a God. What I meant was, the known fact of "shared threat -> people come together" makes the reality of suffering less powerful evidence against the existence of a God.

1UmamiCuboid
Except it really doesn’t, because a truly omni God (omniscient, omnipotent, omnibenevolent, etc) realistically wouldn’t help against the problem of evil. The point of the reality of suffering being evidence against God isn’t about omnibenevolence existing within a vacuum, but about omnibenevolence existing within the context of joint omnipotence and omniscience - Mackie’s inconsistent triad, so to speak. A God possessing both of the latter wouldn’t have to worry about puny human concerns like logic and the like, so logic or rationality-based theodicies (including those around finding a shared enemy) don’t, in my opinion, provide adequate arguments against the problem of evil.

we want a rigorous, formal explanation of exactly how, when, and why you should or should not stick to your precommitment

Well, if we're designing an AI now, then we have the capability to make a binding precommitment, simply by writing code. And we are still in a position where we can hope for the coin to come down heads. So yes, in that privileged position, we should bind the AI to pay up.

However, to the question as stated, "is the decision to give up $100 when you have no real benefit from it, only counterfactual benefit, an example of winning?&q... (read more)

3Wes_W
You're fundamentally failing to address the problem. For one, your examples just plain omit the "Omega is a predictor" part, which is key to the situation. Since Omega is a predictor, there is no distinction between making the decision ahead of time or not. For another, unless you can prove that your proposed alternative doesn't have pathologies just as bad as the Counterfactual Mugging, you're at best back to square one. It's very easy to say "look, just don't do the pathological thing". It's very hard to formalize that into an actual decision theory, without creating new pathologies. I feel obnoxious to keep repeating this, but that is the entire problem in the first place.

We are told no such thing. We are told it's a fair coin and that can only mean that if you divide up worlds by their probability density, you win in half of them. This is defined.

No, take another look:

in the overwhelming measure of the MWI worlds it gives the same outcome. You don't care about a fraction that sees a different result, in all reality the result is that Omega won't even consider giving you $10000, it only asks for your $100.

I wouldn't trust myself to accurately predict the odds of another repetition, so I don't think it would unravel for me. But this comes back to my earlier point that you really need some external motivation, some precommitment, because "I want the 10K" loses its power as soon as the coin comes down tails.

if your decision theory pays up, then if he flips tails, you pay $100 for no possible benefit.

But in the single-shot scenario, after it comes down tails, what motivation does an ideal game theorist have to stick to the decision theory?

Like Parfit's hitchhiker, although in advance you might agree that it's a worthwhile deal, when it comes to the point of actually paying up, your motivation is gone, unless you have bound yourself in some other way.

6Wes_W
That's what the problem is asking! This is a decision-theoretical problem. Nobody cares about it for immediate practical purpose. "Stick to your decision theory, except when you non-rigorously decide not to" isn't a resolution to the problem, any more than "ignore the calculations since they're wrong" was a resolution to the ultraviolet catastrophe. Again, the point of this experiment is that we want a rigorous, formal explanation of exactly how, when, and why you should or should not stick to your precommitment. The original motivation is almost certainly in the context of AI design, where you don't HAVE a human homunculus implementing a decision theory, the agent just is its decision theory.

It should also be possible to milk the scenario for publicity: "Our opponents sold out to the evil plutocrat and passed horrible legislation so he would bankroll them!"

I wish I were more confident that that strategy would actually work...

is the decision to give up $100 when you have no real benefit from it, only counterfactual benefit, an example of winning?

No, it's a clear loss.

The only winning scenario is, "the coin comes down heads and you have an effective commitment to have paid if it came down tails."

By making a binding precommitment, you effectively gamble that the coin will come down heads. If it comes down tails instead, clearly you have lost the gamble. Giving the $100 when you didn't even make the precommitment would just be pointlessly giving away money.

The beggars-and-gods formulation is the same problem.

I don't think so; I think the element of repetition substantially alters it - but in a good way, one that makes it more useful in designing a real-world agent. Because in reality, we want to design decision theories that will solve problems multiple times.

At the point of meeting a beggar, although my prospects of obtaining a gold coin this time around are gone, nonetheless my overall commitment is not meaningless. I can still think, "I want to be the kind of person who gives pennies to beggars, b... (read more)

2Wes_W
This is the point of the thought experiment. Omega is a predictor. His actions aren't just based on what you decide, but on what he predicts that you will decide. If your decision theory says "nah, I'm not paying you" when you aren't given advance warning or repeated trials, then that is a fact about your decision theory even before Omega flips his coin. He flips his coin, gets heads, examines your decision theory, and gives you no money. But if your decision theory pays up, then if he flips tails, you pay $100 for no possible benefit. Neither of these seems entirely satisfactory. Is this a reasonable feature for a decision theory to have? Or is it pathological? If it's pathological, how do we fix it without creating other pathologies?
0hairyfigment
So say it's repeated. Since our observable universe will end someday, there will come a time when the probability of future flips is too low to justify paying if the coin lands tails. Your argument suggests you won't pay, and by assumption Omega knows you won't pay. But then on the previous trial you have no incentive to pay, since you can't fool Omega about your future behavior. This makes it seem like non-payment propagates backward, and you miss out on the whole sequence.

Sorry, but I'm not in the habit of taking one for the quantum superteam. And I don't think that it really helps to solve the problem; it just means that you don't necessarily care so much about winning any more. Not exactly the point.

Plus we are explicitly told that the coin is deterministic and comes down tails in the majority of worlds.

1lolbifrons
If you're not willing to "take one for the team" of superyous, I'm not sure you understand the implications of "every implementation of you is you." It does solve the problem, though, because it's a consistent way to formalize the decision so that on average for things like this you are winning. I think you're missing the point here. Winning in this case is doing the thing that on average nets you the most success for problems of this class, one single instance of it notwithstanding. And this explains why you're missing the point. We are told no such thing. We are told it's a fair coin and that can only mean that if you divide up worlds by their probability density, you win in half of them. This is defined. What seems to be confusing you is that you're told "in this particular problem, for the sake of argument, assume you're in one of the worlds where you lose." It states nothing about those worlds being over represented.

I think that what really does my head in about this problem is, although I may right now be motivated to make a commitment, because of the hope of winning the 10K, nonetheless my commitment cannot rely on that motivation, because when it comes to the crunch, that possibility has evaporated and the associated motivation is gone. I can only make an effective commitment if I have something more persistent - like the suggested $1000 contract with a third party. Without that, I cannot trust my future self to follow through, because the reasons that I would curr... (read more)

1lolbifrons
It seems to me the answer becomes more obvious when you stop imagining the counterfactual you who would have won the $10000, and start imagining the 50% of superpositions of you who are currently winning the $10000 in their respective worlds. Every implementation of you is you, and half of them are winning $10000 as the other half lose $100. Take one for the team.

This is an attempt to examine the consequences of that.

Yes, but if the artificial scenario doesn't reflect anything in the real world, then even if we get the right answer, therefore what? It's like being vaccinated against a fictitious disease; even if you successfully develop the antibodies, what good do they do?

It seems to me that the "beggars and gods" variant mentioned earlier in the comments, where the opportunity repeats itself each day, is actually a more useful study. Sure, it's much more intuitive; it doesn't tie our brains up in kno... (read more)

2Wes_W
Decision theory is an attempt to formalize the human decision process. The point isn't that we really are unsure whether you should leave people to die of thirst, but how we can encode that in an actual decision theory. Like so many discussions on Less Wrong, this implicitly comes back to AI design: an AI needs a decision theory, and that decision theory needs to not have major failure modes, or at least the failure modes should be well-understood. If your AI somehow assigns a nonzero probability to "I will face a massive penalty unless I do this really weird action", that ideally shouldn't derail its entire decision process. The beggars-and-gods formulation is the same problem. "Omega" is just a handy abstraction for "don't focus on how you got into this decision-theoretic situation". Admittedly, this abstraction sometimes obscures the issue.

Perhaps the only appropriate uses for probability 0 and 1 are to refer to logical contradictions (eg P & !P) and tautologies (P -> P), rather than real-world probabilities?

Nope. He's saying that based on his best analysis, it appears to be the case.

Does this particular thought experiment really have any practical application?

I can think of plenty of similar scenarios that are genuinely useful and worth considering, but all of them can be expressed with much simpler and more intuitive scenarios - eg when the offer will/might be repeated, or when you get to choose in advance whether to flip the coin and win 10000/lose 100. But with the scenario as stated - what real phenomenon is there that would reward you for being willing to counterfactually take an otherwise-detrimental action for no reason other than qualifying for the counterfactual reward? Even if we decide the best course of action in this contrived scenario - therefore what?

2Wes_W
Precommitments are used in decision-theoretic problems. Some people have proposed that a good decision theory should take the action that it would have precommitted to, if it had known in advance to do such a thing. This is an attempt to examine the consequences of that.

Also, there is the possibility of future scenarios arising in which Bob could choose to take comparable actions, and we want to encourage him in doing so. I agree that the cases are not exactly analogous.

people who are say, religious, or superstitious, or believe in various other obviously false things

Why do you think you know this?

A while ago, I came across a mathematics problem involving the calculation of the length of one side of a triangle, given the internal angles and the lengths of the other two sides. Eventually, after working through the trigonometry of it (which I have now forgotten, but could re-derive if I had to), I realised that it incorporated Pythagoras' Theorem, but with an extra term based on the cosine of one of the angles. The cosine of 90 degrees is zero, so in a right-angled triangle, this extra term disappears, leaving Pythagoras' Theorem as usual.

The older law that I knew turned out to be a special case of the more general law.

If the hypothetical Omega tells you that they're is indeed a maximum value for happiness, and you will certainly be maximally happy inside the box: do you step into the box then?

This would depend on my level of trust in Omega (why would I believe it? Because Omega said so. Why believe Omega? That depends on how much Omega has demonstrated near-omniscience and honesty). And in the absence of Omega telling me so, I'm rather skeptical of the idea.

1TheOtherDave
For my part, it's difficult for me to imagine a set of observations I could make that would provide sufficient evidence to justify belief in many of the kinds of statements that get tossed around in these sorts of discussions. I generally just assume Omega adjusts my priors directly.

If you believe in G-d then you believe in a being that can change reality just by willing it

OK, so by that definition...if you instead believe in a perfect rationalist that has achieved immortality, lived longer than we can meaningfully express, and now operates technology that is sufficiently advanced to be indistinguishable from magic, including being involved in the formation of planets, then - what label should you use instead of 'G-d'?

0Jiro
I'd probably have to invent a name for it. Or I might use the term "godlike being", implying that the being has some, but not all, characteristics in common with what people think of as God.
4hairyfigment
Khepri Prime, if the sequel to "Worm" goes the way I hope. More seriously, I don't believe any of that, and physics sadly appears to make some of it impossible even in the far future. Most of us would balk at that first word, "perfect," citing logical impossibility results and their relation to idealized induction. So your question makes you seem - let us say disconnected from the discussion. Would you happen to be assuming we reject theism because we see it as low status, and not because there aren't any gods?

I know what a garage would behave like if it contained a benevolent God

Do you, though? What if that God was vastly more intelligent than us; would you understand all of His reasons and agree with all of His policy decisions? Is there not a risk that you would conclude, on balance, "There should be no 'banned products shops'", while a more knowledgeable entity might decide that they are worth keeping open?

3Jiro
If God is more intelligent than me and I don't understand his reasons, that proves too much. It could just as well be that God is evil, and the things that he does that seem good just seem good to me because they are evil on a level that I can't understand.

it greatly changes the "facts" in your "case study".

Actually, does it not add another level of putting Jesus on a pedestal above everyone else?

It changes the equation when comparing Jesus to John Perry (indicating that Jesus' suffering was greatly heroic after all), but perhaps intensifies the "Alas, somehow it seems greater for a hero to have steel skin and godlike powers."

(Btw I'm one of the abovementioned Christians. Just thought I'd point out that the article's point is not greatly changed.)

you cannot be destroyed

In the sense that your mind and magic will hang around, yes. But your material form can still be destroyed, and material destruction of a Horcrux will destroy its ability to anchor the spirit.

So, if two people are mutual Horcruxen, you can still kill person 1, at which point s/he will become a disembodied spirit dependent on person 2, but will cease to be an effective horcrux for person 2. You can then kill person 2, which will permanently kill both of them.

All you really achieve with mutual Horcruxen is to make your Horcrux portable and fragile (subject to illness, aging, accident, etc).

It appears the key issue in creating conflict is that the two groups must not be permitted to get to know each other and become friendly

Because then, of course, they might start attributing each other's negative actions to environmental factors, instead of assuming them to be based on inherent evil.

If those are the unfortunate downsides of policies that are worthwhile overall, then I don't think that qualifies for 'supervillain' status.

I mean, if you're postulating the existence of God, then that also brings up the possibility of an afterlife, etc, so there could well be a bigger picture and higher stakes than threescore years and ten. Sometimes it's rational to say, That is a tragedy, but this course of action is still for the best. Policy debates should not appear one-sided.

If anything, this provides a possible answer to the atheist's question, "Why would God allow suffering?"

0Gradus
"Policy debates should not appear one-sided" doesn't in this case give credence to the idea that a world with suffering implies the possibility of the God. Quite the opposite. That is a post-hoc justification for what should be seen as evidence to lower the probability of "belief in just and benevolent God." This is analogous to EY's example of the absence of sabotage being used as justification for the concentration camps in "Conservation of Expected Evidence"

Isn't this over-generalising?

"religion makes claims, not arguments, and then changes its claims when they become untenable." "claims are all religion has got" "the religious method of claiming is just 'because God said so'"

Which religion(s) are you talking about? I have a hard time accepting that anyone knows enough to talk about all of them.

2Lumifer
Necroing is fine, but you probably shouldn't expect an answer from someone who posted a single comment on LW eight and a half years ago...

The happiness box is an interesting speculation, but it involves an assumption that, in my view, undermines it: "you will be completely happy."

This is assuming that happiness has a maximum, and the best you can do is top up to that maximum. If that were true, then the happiness box might indeed be the peak of existence. But is it true?

3CynicalOptimist
Okay, well let's apply exactly the technique discussed above: If the hypothetical Omega tells you that they're is indeed a maximum value for happiness, and you will certainly be maximally happy inside the box: do you step into the box then? Note: I'm asking that in order to give another example of the technique in action. But still feel free to give a real answer if you'd choose to. Side you didn't answer the question one way or another, I can't apply the second technique here. I can't ask what would have to change in order for you to change your answer.

Email sent about a week ago. Did it get spam-filtered?

I tend to think that the Bible and the Koran are sufficient evidence to draw our attention to the Jehovah and Allah hypotheses, respectively. Each is a substantial work of literature, claiming to have been inspired by direct communication from a higher power, and each has millions of adherents claiming that its teachings have made them better people. That isn't absolute proof, of course, but it sounds to me like enough to privilege the hypotheses.

3entirelyuseless
This is in fact the general problem here. If there is a large group of people claiming that some religion is true, that is quite enough evidence to call your attention to the hypothesis. That is in fact why people's attention is called to the hypothesis: paying attention to what large groups of people say is not remotely close to inventing a random idea.

I think it's pretty clear that animals can feel pain, distress, etc. So we should aim for practices that minimise those things. It's certainly possible - though harder on a mass scale like factory farming.

Also, from a utilitarian perspective, it's clear that eating plants is much more ecologically efficient than feeding plants to animals and then eating the animals. On the other hand, as Elo points out, there are crops and terrain that are not well suited to human food, and might more profitably be used to raise edible animals.

So I'd say that there could b... (read more)

0[anonymous]
Given the lack of predators, we'd have to cull the herbivores anyway.

"if, for example, there was an Islamic theologian who offered to debate the issues with me then I would be inclined to do it and follow where the belief updates lead."

Is that an open offer to theologians of all stripes?

0Alia1d
Yes, theologians of all strips, and philosophers and logicians of all perspectives. As long a they are willing to respond to my questions as well as having me respond to their's. (Though if someone is rude, engages in rhetorical hyperbola, etc. I reserve the right to do those things back to them.) I'll try to check back here to see if anyone wants to do that or e-mail me at alia1dx@gmail.com and I'll give you my private e-mail to carry on a dialog.

In discussing Newcomb's problem, Eliezer at one point stated, "Be careful of this sort of argument, any time you find yourself defining the "winner" as someone other than the agent who is currently smiling from on top of a giant heap of utility."

This aligns well with a New Testament statement from Jesus, "Ye shall know them by their fruits...every good tree bringeth forth good fruit, but a corrupt tree bringeth forth evil fruit."

So, I'm only a novice of the Bayesian Conspiracy, but I can calculate the breast cancer percentages... (read more)

1Personable
So to be converted to a religion (Islam only being an example here) it would have to provide a better moral and positive emotional framework than Christianity? Side-note: This is separate to the question above, but on the topic of the post you provided on positive thinking, I think it may have something to do with religion being less common in a trend among those who are both wealthy and in a higher quality of education (i.e. tertiary education has a lower instance of religion than secondary, public has a lower instance of religion than private, CEOs are less religious than white collar workers, etc) along with a number of other factors that I can't recall, and that these factors do increase tendency towards negative thinking. There is a truth in the stated Christian doctrine (as is shared in Buddhism, Taoism, Hinduism and others) that material goods do not bring happiness or 'salvation'. I personally do not believe, however, that this makes those who are atheists less valid in their beliefs (I would hope so, being an atheist myself).