CuSithBell comments on The Rhythm of Disagreement - Less Wrong

12 Post author: Eliezer_Yudkowsky 01 June 2008 08:18PM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (65)

Sort By: Old

You are viewing a single comment's thread. Show more comments above.

Comment author: CuSithBell 24 May 2012 10:09:34PM 0 points [-]

That's definitely cheating! We don't have access to the means by which X is generated. In the absence of a stated distribution, can we still do better than 50%?

Comment author: cousin_it 25 May 2012 12:10:40AM *  1 point [-]

Well, Thrun's algorithm does better than 50% for every distribution. But no matter what f(x) we choose, there will always be distributions that make the chance arbitrarily close to 50% (say, less than 50%+epsilon). To see why, note that for a given f(x) we can construct a distribution far enough from zero that all values of f(x) are less than epsilon, so the chance of switching prescribed by the algorithm is also less than epsilon.

The next question is whether we can find any other randomized algorithm that does better than 50%+epsilon on any distribution. The answer to that is also no.

1) Note that any randomized algorithm must decide whether to switch or not, based on the contents of the envelope and possibly a random number generator. In other words, it must be described by a function f(x) like in Thrun's algorithm. f(x) doesn't have to be monotonic, but must lie between 0 and 1 inclusive for every x.

2) For every such function f(x), we will construct a distribution of envelopes that makes it do worse than 50%+epsilon.

3) Let's consider for each number x the degenerate distribution D_x that always puts x dollars in one envelope and 2*x in the other.

4) To make the algorithm do worse than 50%+epsilon on distribution D_x, we need the chance of switching at 2*x to be not much lower than the chance of switching at x. Namely, we need the condition f(2*x)>f(x)-epsilon.

5) Now we only need to prove that there exists an x such that f(2*x)>f(x)-epsilon. We will prove that by reductio ad absurdum. If we had f(2*x)≤f(x)-epsilon for every x, we could iterate that and obtain f(x)≥f(x*2^n)+n*epsilon for every x and n, which would make f(x) greater than 1. Contradiction, QED.

Comment author: CuSithBell 25 May 2012 03:21:12AM 0 points [-]

Yes, that all looks sensible. The point I'm trying to get at - the one I think Eliezer was gesturing towards - was that for any f and any epsilon, f(x) - f(2x) < epsilon for almost all x, in the formal sense. The next step is less straightforward - does it then follow that, prior to the selection of x, our expectation for getting the right answer is 50%? This seems to be Eliezer's implication. However, it seems also to rest on an infinite uniform random distribution, which I understand can be problematic. Or have I misunderstood?

Comment author: cousin_it 25 May 2012 08:16:36AM 0 points [-]

That's called an improper prior. Eliezer mentions in the post that it was his first idea, but turned out to be irrelevant to the analysis.

Comment author: CuSithBell 25 May 2012 06:27:19PM 0 points [-]

So I guess we're back to square one, then.

Comment author: cousin_it 25 May 2012 07:11:11PM *  0 points [-]

I don't understand. Which part are you still confused about? To me the whole thing seems quite clear.

Comment author: CuSithBell 25 May 2012 08:00:53PM 0 points [-]

How did Eliezer determine that the expected benefit of the algorithm over random chance is zero?

Comment author: cousin_it 25 May 2012 08:49:23PM *  1 point [-]

He didn't say that, he said the benefit gets closer and closer to zero if you modify the setup in a certain way. I couldn't find an interpretation that makes his statement correct, but at least it's meaningful.

Comment author: CuSithBell 27 May 2012 12:55:08AM 0 points [-]

I don't get why it makes sense to say

the algorithm did make use of prior knowledge about the envelope distribution. (As the density of the differential of the monotonic function, in the vicinity of the actual envelope contents, goes to zero, the expected benefit of the algorithm over random chance, goes to zero.)

without meaning that the expected density of the differential does go to zero - or perhaps would go to zero barring some particular prior knowledge about the envelope distribution. And that doesn't sound like "modifying the setup" to me, that seems like it would make the statement irrelevant. What exactly is the "modification", and what did you decide his statement really means, if you don't mind?

Comment author: cousin_it 27 May 2012 07:55:40AM *  1 point [-]

Sorry, are you familiar with the mathematical concept of limit? Saying that "f(x) goes to zero as x goes to zero" does not imply the nonsensical belief that "x goes to zero".