The goal of this blog-post is to explore the intuition that drives the feeling that the agent who arrives at the conclusion that they should pay Pascal's mugger is being unreasonable, and whether this can be tied in with the analysis of a logical inductor as an algorithm not exploitable by an efficiently computable trader. An analysis of this intuition may help us to design the decision theory of a future Friendly AI so that it is not vulnerable to being Pascal-mugged, which is plausibly a desirable goal from the point of view of obtaining behaviour in alignment with our values.
So, then, consider the following conversation that I recently had with... (read 466 more words →)
I made the following observation to Chris on Facebook which he encouraged me to post here.
My point was basically just that, in reply to the statement "If we don't have such a model to reject, the statement will be tautological", it is in fact true relative to the standard semantics for first-order languages with equality that there is indeed no model-combined-with-an-interpretation-of-the-free-variables for which "x=x" comes out false. That is to say, relative to the standard semantics the formula is indeed a "logical truth" in that sense, although we usually only say "tautology" for formulas that are tautologies in propositional logic (that is, true under every Boolean valuation, a truth-valuation of all subformulas... (read more)