I'm pretty sure that decision theories are not designed on that basis.
You are wrong. In fact, this is a totally standard thing to consider, and "avoid back-chaining defection in games of fixed length" is a known problem, with various known strategies.
Yes, that is the problem in question!
If you want the payoff, you have to be the kind of person who will pay the counterfactual mugger, even once you no longer can benefit from doing so. Is that a reasonable feature for a decision theory to have? It's not clear that it is; it seems strange to pay out, even though the expected value of becoming that kind of person is clearly positive before you see the coin. That's what the counterfactual mugging is about.
If you're asking "why care" rhetorically, and you believe the answer is "you shouldn't be...
Your decision is a result of your decision theory, and your decision theory is a fact about you, not just something that happens in that moment.
You can say - I'm not making the decision ahead of time, I'm waiting until after I see that Omega has flipped tails. In which case, when Omega predicts your behavior ahead of time, he predicts that you won't decide until after the coin flip, resulting in hypothetically refusing to pay given tails, so - although the coin flip hasn't happened yet and could still come up heads - your yet-unmade decision has the same effect as if you had loudly precommitted to it.
You're trying to reason in temporal order, but that doesn't work in the presence of predictors.
You're fundamentally failing to address the problem.
For one, your examples just plain omit the "Omega is a predictor" part, which is key to the situation. Since Omega is a predictor, there is no distinction between making the decision ahead of time or not.
For another, unless you can prove that your proposed alternative doesn't have pathologies just as bad as the Counterfactual Mugging, you're at best back to square one.
It's very easy to say "look, just don't do the pathological thing". It's very hard to formalize that into an actual dec...
But in the single-shot scenario, after it comes down tails, what motivation does an ideal game theorist have to stick to the decision theory?
That's what the problem is asking!
This is a decision-theoretical problem. Nobody cares about it for immediate practical purpose. "Stick to your decision theory, except when you non-rigorously decide not to" isn't a resolution to the problem, any more than "ignore the calculations since they're wrong" was a resolution to the ultraviolet catastrophe.
Again, the point of this experiment is that we w...
There will never be any more 10K; there is no motivation any more to give 100. Following my precommitment, unless it is externally enforced, no longer makes any sense.
This is the point of the thought experiment.
Omega is a predictor. His actions aren't just based on what you decide, but on what he predicts that you will decide.
If your decision theory says "nah, I'm not paying you" when you aren't given advance warning or repeated trials, then that is a fact about your decision theory even before Omega flips his coin. He flips his coin, gets hea...
Decision theory is an attempt to formalize the human decision process. The point isn't that we really are unsure whether you should leave people to die of thirst, but how we can encode that in an actual decision theory. Like so many discussions on Less Wrong, this implicitly comes back to AI design: an AI needs a decision theory, and that decision theory needs to not have major failure modes, or at least the failure modes should be well-understood.
If your AI somehow assigns a nonzero probability to "I will face a massive penalty unless I do this reall...
I'm not sure you've described a different mistake than Eliezer has?
Certainly, a student with a sufficiently incomplete understanding of heat conduction is going to have lots of lines of thought that terminate in question marks. The thesis of the post, as I read it, is that we want to be able to recognize when our thoughts terminate in question marks, rather than assuming we're doing something valid because our words sound like things the professor might say.
OK. I think I see what you are getting at.
First, one could simply reject your conclusion:
However at no point did I do anything that could be described as "simulating you".
The argument here is something like "just because you did the calculations differently doesn't mean your calculations failed to simulate a consciousness". Without a real model of how computation gives rise to consciousness (assuming it does), this is hard to resolve.
Second, one could simply accept it: there are some ways to do a given calculation which are ethical,...
From the point of view of physics, it contains garbage,
But a miracle occurs, and your physics simulation still works accurately for the individual components...?
I get that your assumption of "linear physics" gives you this. But I don't see any reason to believe that physics is "linear" in this very weird sense. In general, when you do calculations with garbage, you get garbage. If I time-evolve a simulation of (my house plus a bomb) for an hour, then remove all the bomb components at the end, I definitely do not get the same result as running a simulation with no bomb.
And apparently insurance companies can make money because the expected utility of buying insurance is lower than it's price.
No, the expected monetary value of insurance is lower than its price. (Assuming that the insurance company's assessment of your risk level is accurate.) You're equivocating between money and utility, which is the source of your confusion.
Suppose I offered a simple wager: we flip a coin, and if it comes up heads, I give you a million dollars. But if it comes up tails, you owe me a million dollars, and I get every cent you earn until...
If I were making music in the style of someone who died six years before I was born, people would probably think I was out of style. I'm not sure if this is the historical fallacy I don't have a name for, where we gloss over differences in a few decades because they're less salient to us than the differences between the 1990s and the 1960s, or if musical styles just change more quickly now.
On the other hand, asteroid mining technologies have some risks of their own, although this only reaches "existential" if somebody starts mining the big ones.
The largest nuclear weapon was the Tsar Bomba: 50 megatonnes of TNT, roughly equivalent to a 3.3-million-tonne impactor. Asteroids larger than this are thought to number in the tens of millions, and at the time of writing only 1.1 million had been provisionally identified. Asteroid shunting at or beyond this scale is by definition a trans-nuclear technology, which means a point comes where the necessary level of trust is unprecedented.
Cromwell's Rule is not EY's invention, and relatively uncontroversial for empirical propositions (as opposed to tautologies or the like).
If you don't accept treating probabilities as beliefs and vice versa, then this whole conversation is just a really long and unnecessarily circuitous way to say "remember that you can be wrong about stuff".
If we're asking what the author "really meant" rather than just what would be correct, it's on record.
...The argument for why zero and one are not probabilities is not, "All objects which are special cases should be cast out of mathematics, so get rid of the real zero because it requires a special case in the field axioms", it is, "ceteris paribus, can we do this without the special case?" and a bit of further intuition about how 0 and 1 are the equivalents of infinite probabilities, where doing our calculations without infinit
What do the following have in common?
You focused on akrasia, and obviously this is a component.
My guess was: they're all wildly underdetermined. "Cheer up" isn't a primitive op. "Don't have sex" or "eat less and exercise more" sound like they might be primitive ops, but can be cashed out in many different ways. "Eat less and exercise more, without excessively disrupting your career/social life/general health/etc" is not a primitive op at all, and may require many non-obvious steps.
I am the downvoter, although another one seems to have found you since. I found your comment to be a mixture of "true, but irrelevant in the context of the quote", and a restatement of non-novel ideas. This is admittedly a harsh standard to apply to a first comment (particularly since you may not have yet even read the other stuff that duplicates your point about human designers being able to avoid local optima!), so I have retracted my downvote.
Welcome to the site, I hope I haven't turned you off.
Good I'm glad we agree on this. Now, why are you trying to defend positions that rely on denying this claim?
I'm not. I entered this discussion mostly to point out that you were equating "corresponds to social behavior" with "does not correspond to anything", which is silly.
It's worse than gender not corresponding to anything. Like in the standard example, it corresponds to multiple things, which don't necessarily agree.
ETA:
Yes, and creeps, or example, want to be treated as a woman with respect to which bathroom they enter.
Do they...
So a man getting an ID card with a typo in the gender field makes him female?
Legally, maybe so, at least until the error is corrected. You'd have to ask a lawyer to be sure.
ID cards are a physical object, which is not determined by biological sex, since as a question of legal fact one can get an ID card of one's self-identified gender if one jumps through the appropriate hoops, even without sex reassignment surgery. (At least that's how it works here in California. I have no idea how it works in other states or countries.)
This seems to me a counterexamp...
So you agree that "gender" as distinct from "sex" doesn't correspond to anything,
I'm pretty sure that ID cards and human interaction are territory, not map. Please don't do the "social constructs basically don't exist" thing, it's very silly.
The discussion of a hypothetical person who wants to change gender (but nothing else) every five minutes is giving me a vibe similar to when someone asks "how does evolution explain a monkey giving birth to a human?" It doesn't. That would falsify the model, much like our hypo...
So instead of every civ fillings its galaxy, we get every civ building one in every galaxy. For this to not result in an Engine on every star, you still have to fine-tune the argument such that new civs are somehow very rare.
There are some hypotheticals where the details are largely irrelevant, and you can back up and say "there are many possibilities of this form, so the unlikeliness of my easy-to-present example isn't the point". "Alien civs exist, but prefer to spread out a lot" does not appear to be such a solution. As such, the requirement for fine-tuning and multiple kinds of exotic physics seem to me like sufficiently burdensome details that this makes a bad candidate.
Implausible premises aside, I'm not convinced this actually resolves the paradox.
The first spacefaring civilization fills the galaxy/universe with Catastrophe Engines at the maximum usable density.
But now the second spacefaring civilization doesn't have any room to build Catastrophe Engines, so they colonize space the regular way. And we're right back at the original problem: either life has to be rare enough that everybody has room to build Engines, or there's lots of life out there that had to expand the non-Engine way but we somehow can't see them.
But... you can already buy many items that are lethal if forcefully shoved down someone's throat. Knives, for example. It's not obvious to me that a lack of lethal drugs is currently preventing anyone from hurting people, especially since many already-legal substances are very dangerous to pour down someone's throat.
From the Overcoming Bias link, "risky buildings" seem to me the clearest example of endangering people other than the buyer.
Interesting. Can I ask you to unpack this statement? I'm curious what exactly you're comparing.
The difference between "has practiced a movement to mastery" and "has never performed a movement before" can be very large, like my powerlifter/snatch example in the other comment. But this is comparing zero practice to a very large amount of practice over a very long period of time. I would find it easy to believe that IQ tests see much greater returns from small amounts of practice.
Also, you bias IQ tests if you repeatedly take them, but you don't do likewise with strength tests so it's much easier to track changes in an individual's strength over time and most anyone whose lifts weights can objectively verify that he has become stronger.
Strength tests are absolutely biased by taking them repeatedly. Athletes call this "specificity".
How do you define/determine this?
The standard definition of strength, which the post cleverly avoided ever stating, is "the ability to produce force against external resistance" or some variant thereof. Force is a well-defined physics term, and can be measured pretty directly in a variety of ways.
Isn't there an "obvious" causal relationship between brain mass and intelligence?
No. Whales aren't smarter than humans.
If by "obvious" you mean "the sort of thing you might guess from first principles", then both are ...
It is interesting, though, how non-general strength is.
There is indeed a widely (unwittingly) held idea that "strength" is a one-dimensional thing: consider, say, superhero comics where the Hulk is stronger than anybody else, which means he's stronger at everything. You never read a comic where the Hulk is stronger at lifting things but Thor is stronger at throwing; that would feel really weird to most people. If the Marvel universe had a comic about strength sports, the Hulk would be the best at every sport.
But this isn't at all how strength wor...
Given that I am wrong, I would prefer being proven wrong to not being proven wrong.
Yours is probably the central case, but "prove me wrong" and "I hope I'm wrong" aren't unheard-of sentiments. For example, a doctor giving a grim diagnosis. I think this can only (?) happen when the (perceived) value on the object level outweighs concerns about ego.
In the above explained situations I would say that in that case simply put their are multiple answers each of which can in the eyes of a different person he true or false.
Yes, except often it really is important to nail down which question we're asking, rather than just accepting that different interpretations yield different answers.
Like he killed a man so its bad BUT that man who was killed had also killed a man so it was good. Choose one it cant be both and the judge of any court knows that.
In logic, we have the law of excluded middle, which stat...
Not all statements are precise enough to be nailed down as definitely true or false. If there's any leeway or ambiguity in exactly what is being stated, there might also be ambiguity in whether it's true or false.
As a trivial example, consider this statement: "If a tree falls in the forest, and there's nobody around to hear it, it doesn't make a sound". Is the statement true or false? Well, it depends on what you mean by "sound": if you mean acoustic vibrations in the air, the tree does make a sound and the statement is false; if you me...
I don't think this is carving reality at the joints.
The free will illusion, at least as presented by Yudkowsky, is that we don't know our own planning algorithm, and understanding how it (probably) works dissolves the illusion, so that "do I have free will" stops even seeming like a question to ask. The illusion is that there was a question at all. The relevant category to watch for is when lots of people want an answer even though nobody can nail down exactly what the question is, or how to tell when you have an answer.
This is a much more specific phenomenon than "elaborate structures", which includes pretty much everything except fundamental particles or the like.
Yes, this is a failure mode of (some forms of?) utilitarianism, but not the specific weirdness I was trying to get at, which was that if you aggregate by min(), then it's completely morally OK to do very bad things to huge numbers of people - in fact, it's no worse than radically improving huge numbers of lives - as long as you avoid affecting the one person who is worst-off. This is a very silly property for a moral system to have.
You can attempt to mitigate this property with too-clever objections, like "aha, but if you kill a happy person, then in ...
"Willpower is not exhaustible" is not necessarily the same claim as "willpower is infallible". If, for example, you have a flat 75% chance of turning down sweets, then avoiding sweets still makes you more likely to not eat them. You're not spending willpower, it's just inherently unreliable.