Pavitra comments on LW is to rationality as AIXI is to intelligence - Less Wrong Discussion
You are viewing a comment permalink. View the original post to see all comments and the full post content.
You are viewing a comment permalink. View the original post to see all comments and the full post content.
Comments (44)
Beware of the representativeness heuristic. Basing your computable approximation on AIXI does not necessarily maximize its accuracy, any more than naive Bayes is inherently superior to its fellow non-Bayesian algorithms due to having "Bayes" in the name.
Using a computable approximation of Solomonoff induction (not AIXI, that's different!) is not some kind of option that can be avoided - modulo some comments about the true razor.
You can warn about its dangers - but we will plunge in anyway.
Ah, I have no idea why I said AIXI. Must have gotten my wires crossed. :|
This seems to leave open the question of what approximation to use, which is essentially the same question posed by the original post. In the real world, for practical purposes, what do you actually use?
Making a computable approximation Solomonoff induction that can be used repeatedly is essentially the same problem as building a stream compressor.
There is quite a bit of existing work on that problem - and it is one of my current projects.
Okay, but what's the actual answer?
I don't understand the question. Can you explain what was wrong with the answer I just gave?
The question is: please recommend a model of rationality that a human can actually use in the real world. It's not clear to me in practice how I would use, say, gzip to help make predictions.
Right, well, the link between forecasting and compression was gone over in this previously-supplied link. See also, the other introductory material on that site:
http://matchingpennies.com/machine_forecasting/
http://matchingpennies.com/introduction/
http://matchingpennies.com/sequence_prediction/
If you want to hear something similar from someone else, perhaps try:
http://www.mattmahoney.net/dc/rationale.html
I understand the theoretical connection. I want a real-world example of how this theoretical result could be applied.
An example of prediction using compression?
E.g. see Dasher. It uses prediction by partial matching.