This is a linkpost for https://keplerlounge.com/posts/entropy/
The Markus Müller paper you link you was interesting. I think he goes a little too wild with it, running into the old trap of quantum immortality / computational immortality. You can't just count future instances of yourself without tracking what actually happens to present instances of yourself - or you can, it's just a bad way to make decisions.
I agree. I don't think he had to attempt to address this problem, if it may be addressed at all.
He has since taken into account the work of experimentalists doing related work, that validates his thesis of Quantum Theory essentially predicting what an observer will see next:
Motivation:
In 1970, Andrey Kolmogorov presented a remarkable vision for the combinatorial foundations of probability theory at the International Congress of Mathematicians in Nice which anticipates the modern development of Artificial Intelligence:
The full text was eventually published in 1983 [1] and Kolmogorov's papers on the subject, where he formulated the fundamental notion of Kolmogorov Complexity, represent a partial fulfilment of his original vision. However, it would be his PhD student Leonid Levin that would complete Kolmogorov's theory of Algorithmic Probability via Levin's Universal Distribution and Levin's Coding theorem which together emphasize the fundamental importance of prefix-free encodings.
As Marvin Minsky, Marcus Hutter and Shane Legg have fully recognised, the implications of Kolmogorov's theory for Artificial Intelligence can't be overstated as it defines the fundamental limits of Machine Learning. In particular, Kolmogorov's theory allows us to reason about Black Swans in AI Safety as Black Swans represent the epistemic limits of a particular ontology.
I wrote this synthesis thanks to the encouragement of Igor Rivin and Alex Kolpakov(aka Sasha), who both recognise that outside a small number of elite institutions such as DeepMind and OpenAI...the current state of the scientific community's understanding of Algorithmic Probability is limited.
Fundamental theorems of Algorithmic Probability:
In this synthesis, I carefully introduce and prove the four fundamental theorems of Algorithmic Probability which lead to Kolmogorov's formulation of Entropy as Expected Kolmogorov Complexity. In summary, these are:
As an application, the last fundamental theorem is then used to derive the Erdős-Euclid theorem which demonstrates that the information content of finitely many primes is insufficient to generate all the integers. There is also a discussion section where I briefly consider the implications of Markus Müller's combinatorial reformulation of the Universal Wave Function [3], where Levin's Universal Distribution predicts the statistical regularities of Physical Reality.
As essential prior knowledge, I recommend reading David McKay's thesis on the correspondence between Machine Learning and Data Compression [4].
Applications:
As a fundamental application of Kolmogorov's definition of Entropy as Expected Kolmogorov Complexity, Sasha and myself have analysed the Game of 20 Questions. It may be demonstrated that, for repeated games, the rank order statistics must satisfy Zipf's Law which may be derived from Levin's Universal Distribution.
If members of the Less Wrong community are interested in a particular application of Algorithmic Probability, feel free to express this interest in the comments below so Sasha and myself may add it to our priority queue. We are particularly inclined to share our analyses of Algorithmic Probability applied to Probabilistic Number Theory as the arguments are simultaneously subtle, verifiable, beautiful and counter-intuitive which creates a structured Gymnasium for the Mind when it comes to reasoning about Machine Epistemology.
As a friend of mine once conjectured, superintelligent machines are bound to follow their own nature but nature may follow a structure if it finds this structure both beautiful and useful.
References: