All of SMK's Comments + Replies

SMK21

I think it would be good if you made clear in the abstract what your contributions to the literature are, and how your results relate to those of e.g. Kierland and Monton (2005).

2glauberdebona
Thanks for the feedback! The summary here and the abstract in the draft paper have been updated; I hope it is clearer now.
SMK10

Yes, of course we can still quibble with the assumptions (like the OP does in some cases), which is why I say "moderate evidence" rather than "completely watertight proof", but given how natural the assumptions are, the evidence is good.

Completeness is arguably not natural (see e.g. Aumann, 1962; Bradley, 2017, §11.5). In particular, I think it is clearly not a requirement of rationality.

SMK*22

Edit: Actually, I think my deeper objection is that most of the critiques here (made by Sammy) are just wrong. For example, of course Dutch books/money pumps frequently get invoked to justify VNM axioms. See for example this.

Sami never mentioned money pumps. And "the Dutch books arguments" are arguments for probabilism and other credal norms[1], not the vNM axioms.

  1. ^

    Again, see Pettigrew (2020) (here is a PDF from Richard's webpage).

3Gustav Alexandrie
I broadly agree with all of Sami's points. However, on this terminological issue I think it is a bit less clear cut. It is true that many decision theorists distinguish between "dutch books" and "money pumps" in the way you are suggesting, and it seems like this is becoming the standard terminology in philosophy.  That said, there are definitely some decision theorists that use "Dutch book arguments" to refer to money pump arguments for VNM axioms. For example, Yaari writes that "an agent that violates Expected Utility Theory is vulnerable to a so-called Dutch book".  Now, given that the entry is called "dutch book theorems" and mostly focuses on probabilism, Sami is still right to point out that it is confusing to say that these arguments support EUT. Maybe I would have put this under "misleading" rather than under "false" though.
SMK7-3

You are conflating the Dutch book arguments for probabilism (Pettigrew, 2020) with the money-pump arguments for the vNM axioms (Gustafsson, 2022). 

1Closed Limelike Curves
We certainly are, which isn't unique to either of us; Savage discusses them all in a single common framework on decision theory, where he develops both sets of ideas jointly. A money pump is just a Dutch book where all the bets happen to be deterministic. I chose to describe things this way because it lets me do a lot more cross-linking within Wikipedia articles on decision theory, which encourages people reading about one to check out the other.
4habryka
I've pretty consistently (by many different people) seen "Dutch Book arguments" used interchangeably with money pumps. My understanding (which is also the SEP's) is that "what is a money pump vs. a dutch book argument" is not particularly well-defined and the structure of the money pump arguments is basically the same as the structure of the dutch book arguments.  This is evident from just the basic definitions:  "A Dutch book is a set of bets that ensures a guaranteed loss, i.e. the gambler will lose money no matter what happens."  Which is of course exactly what a money pump is (where you are the person offering the gambles and therefore make guaranteed money). The money pump Wikipedia article also links to the Dutch book article, and the book/paper I linked describes dutch books as a kind of money pump argument. I have never heard anyone make a principled distinction between a money pump argument and a dutch book argument (and I don't see how you could get one without the other). Indeed, the Oxford Reference says explicitly:  (Edit: It's plausible that for weird historical reasons the exact same argument, when applied to probabilism would be called a "dutch book" and when applied to anything else would be called a "money pump", but I at least haven't seen anyone defend that distinction, and it doesn't seem to follow from any of the definitions)
SMK*53

The normal VNM approach is to start with an agent whose behavior satisfies some common sense conditions: can't be money pumped and so on.

Nitpicks: (1) the vNM theorem is arguably about preference, not choice and behavior; and (2) "can't be money pumped" is not one of the conditions in the theorem.

SMK10

I wrote "I'm really not sure at this point whether UDT is even on the right track" in UDT shows that decision theory is more puzzling than ever which I think you've read? Did you perhaps miss that part?

Yes, missed or forgot about that sentence, sorry.

(BTW this issue/doubt about whether UDT / paying CM is normative for humans is item 1 in the above linked post. Thought I'd point that out since it may not be obvious at first glance.)

Thanks.

Do you have more examples where making such distinctions would be helpful?

I was mostly thinking about discuss... (read more)

SMK30

Here's a related idea that is maybe clearer: Suppose an agent has the ability to self-modify to use any decision theory, would they decide to stick with their current decision theory? (I'm actually not sure what term has been explicitly defined to mean this, so I'll just call it "self-endorsement" for now.)

This sounds similar to what's called "self-recommendation"—see e.g. Skyrms (1982, pp. 707-709), Meacham (2010, §3.3) and Pettigrew (2023). In the abstract Pettigrew writes: "A decision theory is self-recommending if, when you ask it which decision theory... (read more)

SMK*10

Thanks for the clarification!

I do understand from the SEP, like Wei, that sophisticated means "backwards planning", and resolute means "being able to commit to a policy" (correct me if I'm wrong).

That seems roughly correct, but note that there are different interpretations of resolute choice floating around[1], and I think McClennen's (1990) presentation is somewhat unclear at times. Sometimes resoluteness seems to be about the ability to make internal commitments, and other times it seems to be about being sensitive to the dynamic context in a particular ... (read more)

SMK24

I think Sami's comment is entirely fair given the language and framing of the original post. It is of course fine to forget about references, but e.g. "I find it curious that none of my ideas have a following in academia or have been reinvented/rediscovered by academia" and "Clearly academia has some blind spots, but how big?" reads like you don't consider it a possilbity that you might have re-invented something yourself, and that academics are at fault for not taking up your ideas.

7Wei Dai
(It sucks to debate this, but ignoring it might be interpreted as tacit agreement. Maybe I should have considered the risk that something like this would happen and not written my OP.) When I wrote the OP, I was pretty sure that the specific combination of ideas in UDT has not been invented or re-invented or have much of a following in academia, at least as of 2019 when Cheating Death in Damascus was published, because the authors of that paper obviously did a literature search and would have told me if they had found something very similar to UDT in the literature, and I think I also went through the papers it referenced as being related and did not find something that had all of the elements of UDT (that's probably why your references look familiar to me). Plus FDT was apparently considered novel enough that the reviewers of the paper didn't tell the authors that they had to call it by the name of an existing academic decision theory. So it's not that I "don’t consider it a possibility that you might have re-invented something yourself" but that I had good reason to think that's not the case?
SMK*30

I don't think cohesive decision theory is being discussed much, but I'm not sure. Perhaps because the theory is mainly used to argue against the claim that "every decision rule will lead agents who can’t bind themselves to disaster" (p. 20, footnote 34) in the paper, and discussion of its independent interest is relegated to a footnote (footnote 34).

2Daniel Kokotajlo
OK, thanks. So then the mystery remains why academic philosophy isn't more interested in this.
SMK*98

It would be interesting to get an overview of what these are. Or if that's too hard to write down, and there are no ready references, what are your own interests in decision theory?

Yeah, that would be too hard. You might want to look at these SEP entries: Decision Theory, Normative Theories of Rational Choice: Expected Utility, Normative Theories of Rational Choice: Rivals to Expected Utility and Causal Decision Theory. To give an example of what I'm interested in, I think it is really important to take into account unawareness and awareness growth (see §5... (read more)

3Wei Dai
Thanks, will look into your references. I wrote "I'm really not sure at this point whether UDT is even on the right track" in UDT shows that decision theory is more puzzling than ever which I think you've read? Did you perhaps miss that part? (BTW this issue/doubt about whether UDT / paying CM is normative for humans is item 1 in the above linked post. Thought I'd point that out since it may not be obvious at first glance.) Yeah I agree with this to some extent, and try to point out such confusions or make such distinctions when appropriate. (Such as in the CM / indexical values case.) Do you have more examples where making such distinctions would be helpful?
SMK42

There are many many interesting questions in decision theory, and "dimensions" along which decision theories can vary, not just the three usually discussed on LessWrong. It's not clear to me why (i) philosophers should focus on the dimensions you primarily seem to be interested in, and (ii) what is so special about the particular combination you mention (is there some interesting interaction I don't know about maybe?). Furthermore, note that most philosophers probably do not share your intuitions: I'm pretty sure most of them would e.g. pay in counterfactu... (read more)

8Wei Dai
It would be interesting to get an overview of what these are. Or if that's too hard to write down, and there are no ready references, what are your own interests in decision theory? As I mentioned in the previous comment, it happens to solve (or at least seemed like a good step towards solving) a lot of problems I was interested in at the time. Agreed, but my intuitions don't seem so unpopular outside academia or so obviously wrong that there should be so few academic philosophers who do share them. I'm not sure I wouldn't pay either. I see it as more of an interesting puzzle than having a definitive answer. ETA: Although I'm more certain that we should build AIs that do pay. Is that also unclear to you? (If so why might we not want to build such AIs?) Yeah, I'm trying to keep an open mind about that. :) Cool, I'd be interested in any further feedback when you're ready to give them.
SMK60

This is indeed what happens to the best-known decision theories (CDT and EDT): they want to commit to paying, but if they don’t, by the time they get to the Heads world they don’t pay. We call this dynamic instability, because different (temporal) versions of the agent seem to be working against each other.

Unless you are using "dynamic stability" to mean something other than "dynamic consistency", I don't think this is quite right. The standard philosophical theory of dynamic choice, sophisticated choice (see e.g. the SEP entry on decision theory), would not pay but is still dynamically consistent. 

7Wei Dai
Let me check if I understand what you're saying (what these standard academic terms mean). "Dynamic inconsistency" is when you make a plan, initially acts according to that plan, but then later fail to follow the plan. (You failed to take into account what you'd later decide to do when making the plan. This is what the "naive" approach does, according to SEP.) "Dynamic consistency" is when you never do this. (The "sophisticated" approach is to take into account what you'd decide to do later when making the initial plan, so dynamic inconsistency never happens. Also called "backwards reasoning" which I think is the same as "backwards induction" which is a term I'm familiar with.[1]) Rest of this comment assumes this is all correct. I think this is different from what the OP means by "dynamic stability" but perhaps their definition is too vague. (What does "want to commit" mean exactly? Arguably sophisticated choice wants to commit to paying, but also arguably it doesn't because it knows such commitment is futile.) Here's a related idea that is maybe clearer: Suppose an agent has the ability to self-modify to use any decision theory, would they decide to stick with their current decision theory? (I'm actually not sure what term has been explicitly defined to mean this, so I'll just call it "self-endorsement" for now.) (Partly this is about AI, trying to figure out what decision theory future AIs will or should have, but humans also have some ability to choose what decision theory to use, so it seems interesting from that perspective as well.) Sophisticated choice seemingly lacks "self-endorsement" because before it's told how the counterfactual mugger's coin landed, it would self-modify (if it could) into an agent that would pay the counterfactual mugger in the Heads world. (Lots more can be said about this, but I'll stop here and give you a chance to clarify, respond, or ask questions.) 1. ^ I still remember the first moment that I started to question ba
SMK10

The reason for the former is that I (and others) have been unable to find a rigorous formulation of it that doesn't have serious open problems. (I and I guess other decision theory researchers in this community currently think that UDT is more of a relatively promising direction to explore, rather than a good decision theory per se.)

That's fair. But what is it then that you expect academics to engage with? How would you describe this research direction, and why do you think it's interesting and/or important?

6Wei Dai
To quickly recap the history, people on LW noticed some clear issues with "updating" and "physicalist ontology" of the most popular decision theories at the time (CDT/EDT), and thought that switching to "updatelessness" and "logical/algorithmic ontology" was an obvious improvement. (I was the first person to put the two pieces together in an explicit formulation, but they were already being talked about / hinted at in the community.) Initially people were really excited because the resulting decision theories (UDT/FDT) seemed to solve a lot of open problems in one swoop, but then pretty quickly and over time we noticed more and more problems with UDT/FDT that seem to have no clear fixes. So we were initially excited but then increasingly puzzled/confused, and I guess I was expecting at least some academics to follow a similar path, either through engagement with LW ideas (why should they be bothered that much by lack of academic publication?), or from independent invention. Instead academia seems to still be in a state similar to LW when I posted UDT, i.e., the ideas are floating in the air separately and nobody has put them together yet? (Or I guess that was the state of academia before FDT was published in an academic journal, so now the situation is more like some outsiders put the pieces together in a formal publication, but still no academic is following a similar path as us.) I guess it's also possible that academia sort of foresaw or knew all the problems that we'd eventually find with UDT/FDT and that's why they didn't get excited in the first place. I haven't looked into academic DT literature in years, so you're probably more familiar with it. Do you know if they're puzzled/confused by the same problems that we are? Or what are they mostly working on / arguing about these days?
SMK10

Could you perhaps say something about what a Kripkean semantics would look like for your logic?

SMK84

On your first point: as Sami writes, resolute choice is mentioned in the introductory SEP article on dynamic choice (it even has its own section!), as well as in the SEP article on decision theory. And SEP is the first place you go when you want to learn about philosophical topics and find references.

On your second point: as I wrote in my comment above, (i) academics have produced seemingly similar ideas to e.g. updatelessness (well before they were written up on LW) so it is unclear why academics should engage with less rigorous, unpublished proposals tha... (read more)

3Wei Dai
To clarify, by “blind spot” I wasn't complaining that academia isn't engaging specifically with posts written up on LW, but more that nobody in academia seems to think that the combination of "updateless+logical" is clearly the most important or promising direction to explore in decision theory.
SMK*1710

I think the main reason why UDT is not discussed in academia is that it is not a sufficiently rigorous proposal, as well as there not being a published paper on it. Hilary Greaves says the following in this 80k episode:

Then as many of your listeners will know, in the space of AI research, people have been throwing around terms like ‘functional decision theory’ and ‘timeless decision theory’ and ‘updateless decision theory’. I think it’s a lot less clear exactly what these putative alternatives are supposed to be. The literature on those kinds of decision t

... (read more)
2Daniel Kokotajlo
Thanks Sylvester! Yep it looks like cohesive decision theory is basically original UDT.  Do you know what the state of the art is in terms of philosophical critiques of cohesive decision theory? Any good ones? Any good responses to the critiques?
7Wei Dai
The reason for the former is that I (and others) have been unable to find a rigorous formulation of it that doesn't have serious open problems. (I and I guess other decision theory researchers in this community currently think that UDT is more of a relatively promising direction to explore, rather than a good decision theory per se.) And the reason for the latter is the above, plus my personal distaste for writing/publishing academic papers (which I talked about elsewhere in this thread), plus FDT having been published which seems close enough to me. Thank for the references in the rest of your comment. I think I've come across Meacham 2010 and Spohn 2012 before, but forgot about them as I haven't been working actively on decision theory for a while. It does seem that Meacham's cohesive decision theory is equivalent to updateless EDT/CDT. (BTW in The Absent-Minded Driver I referenced a 1997 paper that also has an idea similar to updatelessness, although the authors didn't like it.) On a quick skim of Spohn 2012 I didn't see something that looks like LDT or "algorithmic/logical agent ontology" but it's quite long/dense so I'll take your word on it for now. Still, it seems like none of the academic papers put all of the pieces together in a single decision theory proposal that's equivalent to UDT or FDT? (Please note that UDT as originally described was actually updateless/evidential/logical, not causalist as you wrote in the post that you linked. This has been a historical disagreement between me and Eliezer, where in I leaned towards evidential and he leans towards causal, although these days I just say that I'm confused and don't know what to think.)
SCP114

It may be worth thinking about why proponents of a very popular idea in this community don't know of its academic analogues, despite them having existed since the early 90s[1] and appearing on the introductory SEP page for dynamic choice.

Academics may in turn ask: clearly LessWrong has some blind spots, but how big?

  1. ^

    And it's not like these have been forgotton; e.g., McClennen's (1990) work still gets cited regularly.

SMK10

Thanks.

I am pretty sure they're interchangeable however.

Do you have a reference for this? Or perhaps there is a quick proof that could convince me?

2Yudhister Kumar
Payor's Lemma holds in provability logic, distributivity is invoked when moving from step 1) to step 2) and this can be accomplished by considering all instances of distributivity to be true by axiom & using modus ponens. This section should probably be rewritten with the standard presentation of K to avoid confusion. W.r.t. to this presentation of probabilistic logic, let's see what the analogous generator would be: Axioms: * all tautologies of Christiano's logic * all instances of (x→y)→(□px→□py) (weak distributivity) --- which hold for the reasons in the post Rules of inference: * Necessitation ⟨x,□px⟩ * Modus Ponens ⟨x→y,x,y⟩ Then, again, step 1 to 2 of the proof of the probabilistic payor's lemma is shown by considering the axiom of weak distributivity and using modus ponens. (actually, these are pretty rough thoughts. Unsure what the mapping is to the probabilistic version, and if the axiom schema holds in the same way)
SMKΩ260

You might also find the following cases interesting (with self-locating uncertainty as an additional dimension), from this post.

Sleeping Newcomb-1. Some researchers, led by the infamous superintelligence Omega, are going to put you to sleep. During the two days that your sleep will last, they will briefly wake you up either once or twice, depending on the toss of a biased coin (Heads: once; Tails: twice). After each waking, they will put you back to sleep with a drug that makes you forget that waking. The weight of the coin is determined by what

... (read more)
SMKΩ010

Epistemic Constraint: The probability distribution  which the agent settles on cannot be self-refuting according to the beliefs. It must be a fixed point of : a  such that .

Minor: there might be cases in which there is a fixed point , but where the agent doesn't literally converge or deliberate their way to it, right? (Because you are only looking for  to satisfy the conditions of Brouwer/Kakutani, and not, say, Banach, right?) In other words, it might not always be accurate to say that the agent "set... (read more)

2abramdemski
Yeah, "settles on" here meant however the agent selects beliefs. The epistemic constraint implies that the agent uses exhaustive search or some other procedure guaranteed to produce a fixed point, rather than Banach-style iteration.  Moving to a Banach-like setting will often make the fixed points unique, which takes away the whole idea of FixDT. Moving to a setting where the agent isn't guaranteed to converge would mean we have to re-write the epistemic constraint to be appropriate to that setting.
SMK*Ω7102

A common trope is for magic to work only when you believe in it. For example, in Harry Potter, you can only get to the magical train platform 9 3/4 if you believe that you can pass through the wall to get there.

Are you familiar with Greaves' (2013) epistemic decision theory? These types of cases are precisely the ones she considers, although she is entirely focused on the epistemic side of things. For example (p. 916):

Leap. Bob stands on the brink of a chasm, summoning up the courage to try and leap across it. Confidence helps him in such situations: speci

... (read more)
2abramdemski
Yes, thanks for citing it here! I should have mentioned it, really. I see the Skyrms iterative idea as quite different from the "just take a fixed point" theory I sketch here, although clearly they have something in common. FixDT makes it easier to combine both epistemic and instrumental concerns -- every fixed point obeys the epistemic requirement; and then the choice between them obeys the instrumental requirement. If we iteratively zoom in on a fixed point instead of selecting from the set, this seems harder? If we try the Skyrms iteration thing, maybe the most sensible thing would be to move toward the beliefs of greatest expected utility -- but do so in a setting where epistemic utility emerges naturally from pragmatic concerts (such as A Pragmatists Guide to Epistemic Decision Theory by Ben Levinstein). So the agent is only ever revising its beliefs in pragmatic ways, but we assume enough about the environment that it wants to obey both the epistemic and instrumental constraints? But, possibly, this assumption would just be inconsistent with the sort of decision problem which motivates FixDT (and Greaves).
6SMK
You might also find the following cases interesting (with self-locating uncertainty as an additional dimension), from this post.
SMK*50

We assume two rules of inference:

Necessitation:  
Distributivity: 

Is there a reason why this differs from the standard presentation of K? Normally you would say that K is generated by the following (coupled with substitution):

Axioms:
- All tautologies of propositional logic.
- Distribution: .

Rules of inference:
- Necessitation: .
- Modus ponens: .

3Yudhister Kumar
No particular reason (this is the setup used by Demski in his original probabilistic Payor post). I agree this is nonstandard though! To consider necessitation as a rule of inference & not mentioning modus ponens. Part of the justification is that probabilistic weak distributivity (⊢x→y⟹⊢□px→□py) seems to be much closer to a 'rule of inference' than an axiom for me (or, at least, given the probabilistic logic setup we're using it's already a tautology?). On reflection, this presentation makes more sense to me (or at least gives me a better sense of what's going on / what's different between □p logic and □ logic). I am pretty sure they're interchangeable however.
SMK*Ω010

And, second, the agent will continually implement that plan, even if this makes it locally choose counter-preferentially at some future node.

Nitpick: IIRC, McClennen never talks about counter-preferential choice. Rather, that's Gauthier's (1997) approach to resoluteness.

as devised by Bryan Skyrms and Gerald Rothfus (cf Rothfus 2020b).

Found a typo: it is supposed to be Gerard. (It is also misspelt in the reference list.)

2SCP
Thanks Sylvester; fixed!
SMK130

Some people know that they do not have a solution. Andy Egan, in "Some Counterexamples to Causal Decision Theory" (1999, Philosophical Review)

This should say 2007.

These people all defect in PD and two-box in Newcomb. 

Spohn argues for one-boxing in Reversing 30 years of discussion: why causal decision theorists should one-box.

6Richard_Kennaway
Fixed. Thanks for the reference. I hope this doesn't turn out to be a case of changing CDT and calling the result CDT. ETA: 8 pages in, I'm impressed by the clarity, and it looks like it leads to something reasonably classified as a CDT.
SMK*60

Thanks.

Roughly, you don't actually get to commit your future-self to things. Instead, you just do what you (in expectation) would have committed yourself to given some reconstructed prior.

Agreed.

Just as a literature pointer: If I recall correctly, Chris Meacham's approach in "Binding and Its Consequences" is ultimately to estimate your initial credence function and perform the action from the plan with the highest EU according to that function.

Yes, that's a great paper! (I think we might have had a footnote on cohesive decision theory in a draft of... (read more)

SMK20

What do you mean by "the Bayesian Conditionalization thing" in this context? (Just epistemically speaking, standard Conditionalization is inadequate for dealing with cases of awareness growth. Suppose, for example, that one was aware of propositions {X, Y}, and that this set later expands to {X, Y, Z}. Before this expansion, one had a credence P(X ∨ Y) = 1, meaning Conditionalization recommends remaining certain in X ∨ Y; i.e., one is only permitted to place a credence P(Z) = 0. Maybe you are referring to something like Reverse Bayesianism?)

2Daniel Kokotajlo
I just meant standard conditionalization, which I agree is inadequate for cases of awareness growth. I wasn't making a particularly new point, just commenting aloud as I read along.
SMK10

It's fairly clear to me that the authors do not have any specific and precise method in mind, Bjerring or no Bjerring.

Of course they don't have a specific proposal in the paper. I'm just saying that it seems like they would want to be more precise, or that a full specification requires more work on counterpossibles (which you seem to be arguing against). From the abstract:

While not necessary for considering classic decision theory problems, we note that a full specification of FDT will require a non-trivial theory of logical counterfactuals and algorithmic

... (read more)
2Charlie Steiner
Well, just because something is vague and relies on common sense, doesn't mean you can get whatever answer you want from it.  And there's still plenty of progress to be made in formalizing FDT - it's just that a formalization of an FDT agent isn't going to reference some agent-independent way of computing counterpossibles. Instead it's going to have to contain standards for how best to compute counterpossibles on the fly in response to the needs of the moment.
SMK*11

Right, but it's fairly clear to me that this is not what the authors have in mind. For example, they cite Bjerring (2014), who proposes very specific and precise extensions of the Lewis-Stalnaker semantics.

2Charlie Steiner
It's fairly clear to me that the authors do not have any specific and precise method in mind, Bjerring or no Bjerring. From the paper:   In fact, any agent-independent rule for construction of counterpossibles is doomed, because different questions can cause the same mathematical change to produce different imagined results. What mathematical propositions get chosen to be "upstream" or "downstream" has to depend on what you're thinking of as "doing the changing" or "doing the reacting" for the question at hand. This is important both normatively (e.g. if you were somehow designing an AI that used FDT), and also to understand how humans reason about thought experiments - by constructing the counterfactuals in response to the proposed thought experiment.
SMK10

(See my response to gjm's comment.)

SMK40

To be sure, switching to Bet 1 is great evidence that  is true (that's the whole point), but that's not the sort of reasoning FDT recommends. Rather, the question is if we take the Peano axioms to be downstream of the output of the algorithm in the relevant sense. 

As the authors make clear, FDT is supposed to be "structurally similar" to CDT [1], and in the same way CDT regards the history and the laws to be out of their control in Ahmed's problems, FDT should arguably regard the Peano axioms to be out of their control (i.e., "upstrea... (read more)

2gjm
I think maybe we're running into the problem that FDT isn't (AIUI) really very precisely defined. But I think I agree with Zane's reply to your comment: two (apparently) possible worlds where my algorithm produces different decisions are also worlds where PA proves that it does (or at least they might be; PA can't prove everything that's true)  because those are worlds where I'm running different algorithms. And unless I'm confused (which I very much might be) that's much of the point of FDT: we recognize different decisions as being consequences of running different algorithms.
3Zane
But wouldn't what Peano is capable of proving about your specific algorithm necessarily be "downstream" of the output of that algorithm itself? The Peano axioms are upstream, yes, but what Peano proves about a particular function depends on what that function is.
SMK*40

Regarding Sleeping Counterfact: there seems to be two updates you could make, and thus there should be conceptual space for two interesting ways of being updatelessness in this problem; you could be 'anthropically updateless', i.e., not update on your existence in the standard Thirder way, and you could also be updateless with respect to the researchers asking for money (just as in counterfactual mugging). And it seems like these two variants will make different recommendations.

Suppose you make the first update, but not the second. Then the evidentialist v... (read more)

SMK20

Interesting! Did thinking about those variants make you update your credences in SIA/SSA (or else)?

No, not really! This was mostly just for fun.

SMK52

My follow-up question for almost all of them though, is based on use of the word "should" in the question. Since it presumably is not any moral version of "should", it's presumably a meaning in the direction of "best achieves a desired outcome".

The 'should' only designates what you think epistemic rationality requires of you in the situation. That might be something consequentialist (which is what I think you mean by "best achieves a desired outcome"), like maximizing accuracy[1], but it need not be; you could think there are other norms[2]

To see wh... (read more)

SMK20

Ah, okay, got it. Sorry about the confusion. That description seems right to me, fwiw.

SMK20

Thanks for clarifying. I still don't think this is exactly what people usually mean by ECL, but perhaps it's not super important what words we use. (I think the issue is that your model of the acausal interaction—i.e. a PD with survival on the line—is different to the toy model of ECL I have in my head where cooperation consists in benefitting the values of the other player [without regard for their life per se]. As I understand it, this is essentially the principal model used in the original ECL paper as well.)

3paulfchristiano
The toy model seems like an example, though maybe I misunderstand. I'm just using survival as an example of a thing that someone could care about, and indeed you only have an ECL reason to cooperate if you care about the survival of other agents. I've been using ECL, and understanding others (including the original paper) using ECL to mean: * In a large world there are lots of people whose decisions are correlated with mine. * Conditioned on me doing something that is bad for me and good for someone else, more of those correlated people will do the same. * I will be a beneficiary of many of those decisions---perhaps nearly as often as I pay a cost. * This indirect update dwarfs the direct consequences of my decision. Someone should correct me if this is wrong.
SMK*102

The effective correlation is likely to be (much) larger for someone using UDT.

Could you say more about why you think this? (Or, have you written about this somewhere else?) I think I agree if by "UDT" you mean something like "EDT + updatelessness"[1]; but if you are essentially equating UDT with FDT, I would expect the "correlation"/"logi-causal effect" to be pretty minor in practice due to the apparent brittleness of "logical causation".

Correlation and kindness also have an important nonlinear interaction, which is often discussed under the heading of “ev

... (read more)
4paulfchristiano
Yeah, by UDT I mean an updateless version of EDT. I'm not sure I understand the distinction (or what you mean "your opponent's universe"), and I might have miscommunicated. What I mean is: * Correlation in a large world means that conditioned on you cooperating, a lot of people cooperate. * What matters is how much you care about the welfare of the people who cooperate relative to the welfare of the beneficiaries. * So you can end up cooperating owing to a combination of correlation+kindness even if you are neither correlated with nor care about your opponent.
SMK70

Related: A bargaining-theoretic approach to moral uncertainty by Greaves and Cotton-Barratt. Section 6 is especially interesting where they highlight a problem with the Nash approach; namely that the NBS is variant to whether (sub-)agents are bargaining over all decision problems (which they are currently facing and think they will face with nonzero probability) simultaneously, or whether all bargaining problems are treated separately and you find the solution for each individual problem—one at a time.

In the 'grand-world' model, (sub-)agents can bargain ac... (read more)

SMK60

Now, let's pretend you are an egalitarian. You still want to satisfy everyone's goals, and so you go behind the veil of ignorance, and forget who you are. The difference is that now you are not trying to maximize expected expected utility, and instead are trying to maximize worst-case expected utility.

Nitpick: I think this is a somewhat controversial and nonstandard definition of egalitarianism. Rather, this is the decision theory underlying Rawls' 'justice as fairness'; and, yes, Rawls claimed that his theory was egalitarian (if I remember correctly), but... (read more)

SMK*30

I agree that the latter two examples have Moorean vibes, but I don't think they strictly speaking can be classified as such (especially the last one). (Perhaps you are not saying this?) They could just be understood as instances of modus tollens, where the irrationality is not that they recognize that their belief has a non-epistemic generator, but rather that they have an absurdly high credence in , i.e. "my parents wouldn't be wrong" and "philosophers could/should not be out of jobs".

SMK*11

The same holds if Alice is confident in Bob's relevant conditional behavior for some other reason, but can't literally view Bob's source code. Alice evaluates counterfactuals based on "how would Bob behave if I do X? what about if I do Y?", since those are the differences that can affect utility; knowing the details of Bob's algorithm doesn't matter if those details are screened off by Bob's functional behavior.

Hm. What kind of dependence is involved here? Doesn't seem like a case of subjunctive dependence as defined in the FDT papers; the two algorithms a... (read more)

SMK*10

I'm not claiming this (again, it's about relative not absolute likelihood).

I'm confused. I was comparing the likelihood of (3) to the likelihood of (1) and (2); i.e. saying something about relative likelihood, no?

I'm not saying this is likely, just that this is the most plausible path I see by which UDT leads to nice things for us.

I meant for my main argument to be directed at the claim of relative likelihood; sorry if that was not clear. So I guess my question is: do you think the updatelessness-based trade you described is the most plausible type of acau... (read more)

SMK41

I had something like the following in mind: you are playing the PD against someone implementing "AlienDT" which you know nothing about except that (i) it's a completely different algorithm to the one you are implementing, and (ii) that it nonetheless outputs the same action/policy as the algorithm you are implementing with some high probability (say 0.9), in a given decision problem.

It seems to me that you should definitely cooperate in this case, but I have no idea how logi-causalist decision theories are supposed to arrive at that conclusion (if at all).

4Rob Bensinger
This is why I suggested naming FDT "functional decision theory" rather than "algorithmic decision theory", when MIRI was discussing names. Suppose that Alice is an LDT Agent and Bob is an Alien Agent. The two swap source code. If Alice can verify that Bob (on the input "Alice's source code") behaves the same as Alice in the PD, then Alice will cooperate. This is because Alice sees that the two possibilities are (C,C) and (D,D), and the former has higher utility. The same holds if Alice is confident in Bob's relevant conditional behavior for some other reason, but can't literally view Bob's source code. Alice evaluates counterfactuals based on "how would Bob behave if I do X? what about if I do Y?", since those are the differences that can affect utility; knowing the details of Bob's algorithm doesn't matter if those details are screened off by Bob's functional behavior.
SMK*20

What's your take on playing a PD against someone who is implementing a different decision algorithm to the one you are implementing, albeit strongly (logically) correlated in terms of outputs?

2JBlack
I'm guessing that "strongly logically correlated in terms of outputs" means that it has the same outputs for a large fraction of inputs (but not all) according to some measure over the space of all possible inputs. If that's all you know, then there will likely be nearly zero logical correlation between your outputs for this instance, and what you will decide depends upon what your decision algorithm does when there is close to zero logical correlation. If you have more specific information than just existence of a strong logical correlation in general, then you should use it. For example, you may be told that the measure over which the correlation is taken is heavily weighted toward your specific inputs for this instance, and that the other player is given the same inputs. That raises the logical correlation between outputs for this instance, and (if your decision algorithm depends upon such things) you should cooperate.
2Rob Bensinger
Depends on the decision algorithm! Do you have a specific one in mind? E.g., LDT will defect against CDT.
SMK*152

Insofar as I have hope in decision theory leading us to have nice things, it mostly comes via the possibility that a fully-fleshed-out version of UDT would recommend updating "all the way back" to a point where there's uncertainty about which agent you are. (I haven't thought about this much and this could be crazy.)

This was surprising to me. For one, that seems like way too much updatelessness. Do you have in mind an agent self-modifying into something like that? If so, when and why? Plausibly this would be after the point of the agent knowing whether it ... (read more)

2Richard_Ngo
Yepp, I agree. I'm not saying this is likely, just that this is the most plausible path I see by which UDT leads to nice things for us. I'm not claiming this (again, it's about relative not absolute likelihood).  
SMK*30

From Arif Ahmed's Evidence, Decision and Causality (ch. 5.4, p. 142-143; links mine):

Deliberating agents should take their choice to be between worlds that differ over the past as well as over the future. In particular, they differ over the effects of the present choice but also over its unknown causes. Typically these past differences will be microphysical differences that don’t matter to anyone. But in Betting on the Past they matter to Alice.

. . .

On this new picture, which arises naturally from [evidential decision theory]. . ., it is misleading t

... (read more)
SMK*10

The difficulty is in how to weight the frequency/importance of the situations they face. 

I agree with this. On the one hand, you could just have a bunch of Procreation problems which would lead to the FDTer ending up with a smaller pot of money; or you could of course have a lot of Counterfactual Muggings in which case the FDTer would come out on top—at least in the limit.

SMK30

Ah, nice. I was just about to recommend sections 2.6.2 and 3 of Multiverse-wide Cooperation via Correlated Decision Making by Caspar.

1Martín Soto
Nice, thank you! I will delve into that one as well when I have the time :-)
SMK*32

From this, they trivially conclude that EDT will have higher stakes than CDT: if there are more Good Twins (Evil Twins), EDT will recommend one-boxing (two-boxing) very strongly, since this will provide evidence to you about many agents doing the same. But I'm not satisfied with this answer, because if you don't know whether more Good Twins or Evil Twins exist, you won't be obtaining that evidence (upon taking the decision)!

I don't think this is a situation of evidential symmetry which would warrant a uniform distribution (i.e. you can't just say that "you... (read more)

3Martín Soto
Thank you for your comment, Sylvester! As it turns out, you're right! Yesterday I discussed this issue with Caspar Oesterheld (one of the authors). Indeed, his answer to this objection is that they believe there probably are more positively than negatively correlated agents. Some arguments for that are evolutionary pressures and the correlation between decision theory and values you mention. In this post, I was implicitly relying on digital minds being crazy enough as for a big fraction of them to be negatively correlated to us. This could plausibly be the case in extortion/malevolent actors scenarios, but I don't have any arguments for that being probable enough. In fact, I had already come up with a different objection to my argument. And the concept of negatively correlated agents is generally problematic for other reasons. I'll write another post presenting these and other considerations when I have the time (probably the end of this month). I'll also go over Greaves [2016], thank you for that resource!
SMK*10

Procreation* gives both FDT and CDT agents (and indeed, all agents) the same dilemma. FDT agents procreate and live miserably; CDT agents don't procreate and almost certainly don't exist. FDT beats CDT in this dilemma.

This doesn't seem right: you already exist! In order to say that "FDT beats CDT" I think you have to argue that one should care about the number of branches you exist in—which is what you plausibly have uncertainty about, not about whether this very instance of you exists. (And this is arguably just about preferences, as Christiano writes abo... (read more)

1Heighn
Thanks for your reaction! Sure, I already exist; together with the fact that I make the exact same decision my father made, that implies I procreate and therefore I'm not a CDT'er. The point with these problems is, I believe, that your decision procedure is implemented at least 1 time, but possibly 2 times throughout time - depending on what your decision procedure outputs. In Procreation*, if "my" decision procedure outputs "procreate", it first does so "in" my father, who then procreates, causing me to exist. I then also procreate. But if "my" decision procedure outputs "don't procreate", it also first does so "in" my father, who then doesn't procreate, and then I don't exist.  The question "Should I procreate?" is a bit misleading, then, as I possibly don't exist. Or, we indeed assume I do exist; but then it's not much of a decision problem anymore. If I exist, then my father procreated, and I necessarily procreate too.
Load More