That seems a little bit ad hoc to me. Either you care about dogs (and then even the tiniest non-zero amount of caring should be enough for the argument) or you don't.
Eliezer handled this sort of objection in Newcomb's Problem and Regret of Rationality:
You can't tell me, first, that above all I must conform to a particular ritual of cognition, and then that, if I conform to that ritual, I must change my morality to avoid being Dutch-booked. Toss out the losing ritual; don't change the definition of winning.
What you are doing here is insisting that I conform to your ritual of cognition (i.e. total utilitarianism with real-number valuation and additive aggregation). I see no reason to accede to such a demand.
The following are facts about what I do and don't care about:
1) All else being equal, I prefer that a dog not be tortured.
2) All else being equal, I prefer that my grandmother not be tortured.
3) I prefer any number of dogs being tortured to my grandmother being tortured.
4 through ∞) Some other stuff about my preferences, skipped for brevity.
#s 2 and 3 are very strong preferences. #1 is less so.
Now I want to find a moral calculus that captures those facts. You, on the other hand, are telling me that, first, I must accept your moral calculus, and then, that if I do so, I must toss out one of the aforementioned preferences.
I decline to do either of those things. (As Eliezer says in the above link: The utility function is not up for grabs.)
But for the sake of the argument : if a hyperintelligent alien came to the Earth (eg. an AI), what would you propose how the alien should figure out which mechanisms in the universe should be of moral concern?
I don't know. This is the kind of thing that demonstrates why we need FAI theory and CEV.
What would you think of the agent’s morality if it discounted your welfare lexically?
I would think that its morality is different from mine. Also, I would be sad, because presumably such a morality on the AI's part would result in bad things for me. Your point?
Subscribe to RSS Feed
= f037147d6e6c911a85753b9abdedda8d)
Let's look at a slightly more complicated example:
A patient comes in to the hospital because he's sick, his vitals are taken (C), and based on C, a doctor prescribes medication A. Sometime later, the patient dies (Y). Say this happens for lots of patients, and we form an empirical distribution p(C,A,Y) from these cases. Things we may want to represent:
marginal probability of death: p(Y)
the policy the doctor followed when prescribing the medicine: p(A|C)
the probability that someone would die given that they were given the medicine: p(Y|A)
the "causal effect" of medicine on death: ???
The issue with the "causal effect" is that the doctor's policy, if it is sensible, will be more likely to prescribe A to people who are already very sick, and not prescribe A to people who are mostly healthy. Thus it may very well turn out that p(Y|A) is higher than the probability p(Y) (this happens for example with HIV drugs). But surely this doesn't mean that medicine doesn't help! What we need is the probability of death given that the doctor arbitrarily decided to give medicine, regardless of background status C. This arbitrary decision "decouples" the influence of the patient's health status and the influence of the medicine in the sense that if we average over health status for patients after such a decision was made, we would get just the effect of the medicine itself.
For this we need to distinguish an "arbitrary" decision to give medicine, divorced from the policy. Call this arbitrary decision A* . You may then claim that we can just use rules of conditional probabilities on events Y,C,A, and A* , and we do not need to involve anything else. Of course, if you try to write down sensible axioms that relate A* with Y,C,A you will notice that essentially you are writing down standard axioms of potential outcomes (one way of thinking about potential outcomes is they are random variables after a hypothetical "arbitrary decision" like A* ). For example, if were to arbitrarily decide to prescribe medicine just precisely for the patients the doctor prescribed medicine to, we would get that p(Y|A) = p(Y|A* ) (this is known as the consistency axiom). You will find that your A* functions as an intervention do(a).
There is a big literature on how intervention events differ from observation events (they have different axiomatizations, for example). You may choose to use a notation for interventions that does not look different from observations, but the underlying math will be different if you wish to be sensible. That is the point. You need new math beyond math about evidence (which is what probability theory is).
It seems to me that we often treat EDT decisions with some sort of hindsight bias. For instance, given that we know that the action A (turning on sprinklers) doesn't increase the probability of the outcome O (rain) it looks very foolish to do A. Likewise, a DT that suggests doing A may look foolish. But isn't the point here that the deciding agent doesn't know that? All he knows is, that P(E|A)>P(E) and P(O|E)>P(O). Of course A still might have no or even a negative causal effect on O, but yet we have more reason the believe otherwise. To illustrate that, consider the following scenario:
Imagine you find yourself in a white room with one red button. You have no idea why you're there and what this is all about. During the first half hour you are undecided whether you should press the button. Finally your curiosity dominates other considerations and you press the button. Immediately you feel a blissful release of happiness hormones. If you use induction it seems plausible to infer that considering certain time intervalls (f.i. of 1 minute) P(bliss|button) > P(bliss). Now the effect has ceased and you wished to be shot up again. Is it now rational to press the button a second time? I would say yes. And I don't think that this is controversial. And since we can repeat the pattern with further variables it should also work with the example above.
From that point of view it doesn't seem to be foolish at all - talking about the sprinkler again - to have a non-zero credence in A (turning sprinkler on) increasing the probability of O (rain). In situations with that few knowledge and no further counter-evidence (which f.i. might suggest that A might have no or a negative influence on O) this should lead an agent to do A.
Considering the doctor, again, I think we have to stay clear about what the doctor actually knows. Let's imagine a doctor who lost all his knowledge about medicine. Now he reads one study which shows that P(Y|A) > P(Y). It seems to me that given that piece of information (and only that!) a rational doctor shouldn't do A. However, once he reads the next study he can figure out that C (the trait "cancer") is confounding the previous assessment because most patients who are treated with A show C as well, whereas ~A don't. This update (depending on the probability values respectively) will then lead to a shift favoring the action A again.
To summarize: I think many objections against EDT fail once we really clarify what the agent knows respectively. In scenarios with few knowledge EDT seems to give the right answers. Once we add further knowledge an EDT updates his beliefs and won’t turn on the sprinkler in order to increase the probability of rain. As we know it from the hindsight bias it might be difficult to really imagine what actually would be different if we didn’t know what we do know now.
Maybe that's all streaked with flaws, so if you find some please hand me over the lottery tickets ; )