I've recently been thinking about Solomonoff induction, and in particular the free choice of Universal Turing Machine.
One variable that seems like a potential choice here is a human brain (my brain for example). It's obviously a bit of a weird choice, but I don't see any reason to disprefer it over a Python interpreter, given that the whole point of Solomonoff induction is to define a prior and so my knowledge of physics or atoms shouldn't really come into play when choosing a UTM.
Concretely, the UTM would be a human simulated in an empty room with an infinitely large notebook symbolizing the tape. It's output would be an encoding of sensory data, and it's input would be a string of instructions in english.
If we do that, then the sentence "the woman at the end of the street is a witch, she did it" suddenly becomes one of the simplest hypotheses that are available to us. Since the english sentence is so short, we now basically just need to give an encoding of who that woman is and what the action in question is, which is probably also going to be a lot shorter in human language than machine language (since our UTM already understands basic physics, society, etc.), and then our simulated human (which since Solomonoff induction doesn't have runtime constraints can take as much time as they want) should be able to produce a prediction of historical sensory input quite well, with relatively little input.
I feel like I must be missing something in my understanding of Solomonoff induction. I have a lot more thoughts, but maybe someone else has already thought of this and can help me understand this. Some thoughts that come to mind:
- I don't know how to build a human brain, but I know how to build a machine that runs a Python interpreter. In that sense I understand a Python interpreter a lot better than I do a human brain, and using it as the basis of Solomonoff induction is more enlightening
- There is a weird circularity about choosing a human brain (or your own brain in particular) as the UTC in Solomonoff induction that I can't quite put my finger on
- Maybe I am misunderstanding the Solomonoff induction formalism so that this whole construction doesn't make any sense
I see how the human brain based computer would give an advantage in encoding a video feed. I still don't see how this relates to the witch hypothesis. You still have to say what the witch did and why that got you the video feed you got, right? The witch hypothesis will most likely only beat the best possible causal model if there actually is magic going on (though, there is a problem in that you might not know the best possible causal model).
If the human can execute arbitrary programs and is computable, and can interpret messages of the form "run this program on the rest of the message", then by definition the human brain based computer is a UTM, so it can be used in Solomonoff induction, weirdly enough. However, there is a concern in that the UTM is meant to be a prior, whereas the brain is more of a representation of a posterior. So it will be able to overfit things you already know. This might not be a problem if you were using CDT but would be a problem for UDT, since UDT isn't supposed to change its prior as it gets more observations (this would make it stop caring about non-actual worlds in e.g. counterfactual mugging, leading to dynamic inconsistency).
In general if you're using UDT then your choice of prior is a choice of which possible worlds you care about and how much. There won't be universally compelling arguments for a particular prior the same way there aren't universally compelling arguments for particular values.
(also, you're welcome, this was useful to think about from my end too!)
I think the last paragraph was the most clarifying to me in the exchange so far. If you would be up for it, I think it would be great if you could edit your top-level comment to include that paragraph and maybe also some of the other things said in this thread (though obviously no obligation, just seems better for future people who might have a similar question, to have everything in one top-level place).