Tim_Freeman
Tim_Freeman has not written any posts yet.

Tim_Freeman has not written any posts yet.

Benja --
I disagree with Tyrrell (see below), but I can give a version of Tyrrell's "trivial" formalization:
We want to show that:
Averaging over all theories T, P(T makes correct predictions | T passes 10 tests) > P(T makes correct predictions)
By Bayes' rule,
P(T makes correct predictions | T passes 10 tests) = P(T makes correct predictions)
So our conclusion is equivalent to:
Averaging over all theories T, P(T passes 10 tests | T makes correct predictions) / P(T passes 10 tests)
1
which is equivalent to
Averaging over all theories T, P(T passes 10 tests | T makes correct predictions) > P(T passes 10 tests)
which has to be true for any plausible definition... (read more)
Three points in response to Eliezer's post and one of his replies:
* A limited time horizon works better than he says. If an AI wants to put its world into a state desired by humans, and it knows that the humans don't want to live in a galaxy that will be explode in a year, then an AI that closes its books in 1000 years will make sure that the galaxy won't explode one year later.
* An unbounded utility works worse than he says. Recall the ^^^^ operator originally by Knuth (see http://en.wikipedia.org/wiki/Knuth%27s_up-arrow_notation) that was used in the Pascal's Mugging article at http://lesswrong.com/lw/kd/pascals_mugging_tiny_probabilities_of_vast/.
If one allows unbounded utilities, then one has allowed a utility of about 3^^^^3 that has no low-entropy representation.... (read more)
Hugo Mercier's citation above for "Believe it or Not" by Hasson et al. wants money to give you the article. The article is available for free from Hasson's home page at:
http://home.uchicago.edu/~uhasson/
The direct URL is:
The premise is that a rational agent would start out convinced that this story about the alien that knows in advance what they'll decide appears to be false.
The Kolomogorov complexity of the story about the alien is very large because we have to hypothesize some mechanism by which it can extrapolate the contents of minds. Even if I saw the alien land a million times and watched the box-picking connect with the box contents as they're supposed to, it is simpler to assume that the boxes are some stage magic trick, or even that they are an exception to the usual laws of physics.
Once we've done enough experiments that we're forced... (read more)
>And now the philosopher comes and presents their "thought experiment" - setting up a scenario in which, by
>stipulation, the only possible way to save five innocent lives is to murder one innocent person, and this murder is
>certain to save the five lives. "There's a train heading to run over five innocent people, who you can't possibly
>warn to jump out of the way, but you can push one innocent person into the path of the train, which will stop the
>train. These are your only options; what do you do?"
If you are looking out for yourself, it's an easy decision, at least in the United States. There... (read more)