Epistemic status: too good to be true? Please check my math.

We've known for a while that Löb's theorem fails when proof is relaxed to probabilistic belief. This has pros and cons. On the pro side, it means there's no Löbian Obstacle to probabilistic self-trust. On the con side, it means that some Löb-derived insights for proof-based decision theory don't translate to probabilistic decision theory, at least not as directly as one might hope. In particular, it appeared to dash hopes for probabilistic generalizations of the "Löbian handshake" for cooperation.

Recently, Andrew Critch wrote about the Payor Lemma, which allows for a very similar "modal handshake" without Löb's Theorem. The lemma was proved using the same modal assumptions as Löb's, so on the surface it may appear to be just a different method to achieve similar results, whose main advantage is that it is much easier to prove (and therefore explain and understand) than Löb's Theorem.

But, a natural question arises: does Payor's Lemma have a suitable probabilistic version?

I'll give an affirmative proof; but I haven't confirmed that the assumptions are reasonable to my satisfaction.

Setup

Let  be a language in first-order logic, expressive enough to represent its sentences  as quoted terms , eg, through Gödel numbering; and with a probability function symbol on these terms, , which can be equated with (some representation of) rational numbers, e.g. , etc. I also assume the system can reason about these rational numbers in the basic ways you'd expect.

For all  and all , we have:

  • If  , then  .
  • If  , then  .

(These assumptions might look pretty minimal, but they aren't going to be true for every theory of self-referential probability; more on this later.)

Let  abbreviate the sentence  for any  and some globally fixed constant  strictly between 0 and 1. This is our modal operator. (This modal operator is often called "p-belief" with the constant  named  instead, but I wanted to reserve  for the probability operator.)

Some important properties of :

Necessitation. If , then , for any .

Proof: Since  implies , and , we have  , which is to say, . [End proof.]

Weak distrubitivity. If , then .

Proof: When , we have , so . [End proof.]

(Regular distributivity would say  implies . The assumption  is stronger than , so the above is a weaker form of distributivity.)

Theorem Statement

  • If   ,  then   .

Proof

  1. , by tautology .
  2. So , from 1 by weak distributivity.
  3. Suppose .
  4.  from 2 and 3, by propositional logic.
  5.  from 4 by necessitation.
  6.  from 5 and 3.
    [End proof.]

Discussion

Comparison to Original Proof

The proof steps mirror Critch's treatment very closely. The key difference is step 2, IE, how I obtain a statement like . Critch uses distributivity, which is not available to me. Here's what I could get instead:[1]

  • ?
    • Suppose , ie, .
    • Rewrite .
    • Now suppose , that is, .
    • .
    • .
    • .
    • .
    • .
    • So we only get:
      • ,
      • where  abbreviates  and we have .

So, it seems attempted applications of distributivity will typically create weakened belief operators, which would get in the way of the proof (very similar to how probabilistic Löb fails).

However, the specific application we want happens to go through, due to a logical relationship between  and ; namely, that  is a weaker statement than .

This reveals a way in which the assumptions for Payor's Lemma are importantly weaker than those required for Löb to go through.

So, the key observation I'm making is that weak distributivity is all that's needed for Payor, and it seems much more plausible for self-referential probabilistic reasoning than regular distributivity.

Reasonable Axioms?

This proof shouldn't be regarded as a final result, since I'm not sure the probability axioms I gave are good. They're consistent: we can consistently assign probability 1 to everything, at least. But obviously that's not an intended model. We can fix this by adding the axiom . I excluded this simply because I don't need it in the proof. If we want to do full probability, we will also want to add something like the inclusion-exclusion principle. So our four axioms could be:

0. 

1.

2.  

3. 

These are similar to the axioms for logical-probability in From Classical to Intuitionistic Probability

But is the resulting system consistent? While the axioms may look extremely innocuous, the fact that they embed information about the probability distribution within the same logic which the probability distribution has beliefs over creates a possibly dangerous amount of self-reference. We need to worry about self-referential sentences like the "probabilistic Liar sentence" 

We can derive something like the reflection schema in Definability of Truth in Probabilistic Logic, which is reassuring since stuff like  is consistent in that system:

  • If , then .

This is true since  implies  for any . But we can similarly derive a principle analogous to one rejected in that paper:

  • If , then .

The presence of the  changes things enough that the relationship to the original paper isn't obvious.

The key difference between the kind of self-reference I need and the kind explored in Definability of Truth in Probabilistic Logic is that that paper focused on self-knowledge about the probability values themselves. What I need is self-knowledge about the rules of probability followed. Unfortunately, the paper doesn't address this, so more work is required to check whether my two axioms could be added consistently.

My hope is that the proof of Payor's Lemma I suggest shouldn't depend on too much esoteric stuff about how exactly we define self-referential probability and make it consistent. So, my conjecture would be that for many approaches to self-referential probability, some translation of Payor's Lemma works.

But my theorem above is only very limited evidence of this.

  1. ^

    I'm not being super careful about what axioms I need, below; I'm assuming something like "ordinary probabilistic reasoning". I'm not proving that I can't have full distributivity (which would require a different approach); I'm just illustrating why it doesn't seem to make sense for probabilistic beliefs.

New Comment
7 comments, sorted by Click to highlight new comments since:

The lemma was proved using the same modal assumptions as Löb's

I think the lemma doesn't need internal necessitation (). Though it's still referenced in a proof of non-exploitability.

I think that step 6 is supposed to say "from 5 and 3" instead of "from 4 and 1"?

Thanks, fixing!

Nice!

⊬a

Should be , right?

Thanks, fixing!

  • .
  • .

Are you deriving these two points only using the axioms provided in the "Setup" section, or are you assuming all four axioms stated afterward here?

Ah, good point. I was using some other axioms. I'll clarify.