You're looking at Less Wrong's discussion board. This includes all posts, including those that haven't been promoted to the front page yet. For more information, see About Less Wrong.

FrameBenignly comments on Open thread, Dec. 21 - Dec. 27, 2015 - Less Wrong Discussion

2 Post author: MrMind 21 December 2015 07:56AM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (230)

You are viewing a single comment's thread. Show more comments above.

Comment author: FrameBenignly 22 December 2015 07:19:00AM -1 points [-]

I tried to get a discussion going on this exact subject in my post this week, but there seemed to be little interest. A major weakness of the standard Bayesian inference method is that it assumes a problem only has two possible solutions. Many problems involve many possible solutions, and many times the number of possible solutions is unknown, and in many cases the correct solution hasn't been thought of yet. In such instances, confirmation through inductive inference may not be the best way of looking at the problem.

Comment author: IlyaShpitser 22 December 2015 03:19:21PM *  5 points [-]

A major weakness

Where did you get this from? Maintaining beliefs over an entire space of possible solutions is a strength of the Bayesian approach. Please don't talk about Bayesian inference after reading a single thing about updating beliefs on whether a coin is fair or not. That's just a simple tutorial example.

Comment author: FrameBenignly 22 December 2015 04:03:19PM *  0 points [-]

If I have 3 options, A, B, and C, and I'm 40% certain the best option is A, 30% certain the best option is B, and 30% certain the best option is C, would it be correct to say that I've confirmed option A instead of say my best evidence suggests A? This can sort of be corrected for with the standard Bayesian confirmation model, but the problem becomes larger as the number of possibilities increases to the point where you can't get a good read on your own certainty, or to the point where the number of possibilities is unknown.

Comment author: IlyaShpitser 22 December 2015 04:32:36PM 1 point [-]

I don't understand your question. Is this about maintaining beliefs over hypotheses or decision-making?

Comment author: FrameBenignly 22 December 2015 04:56:53PM 0 points [-]

I'm arguing that Bayesian confirmation theory as a philosophy was originally conceived as a model using only two possibilities (A and ~A), and then this model was extrapolated into problems with more than two possibilities. If it had been originally conceived using more than two possibilities, it wouldn't have made any sense to use the word confirmation. So explanations of Bayesian confirmation theory will often entail considering theories or decisions in isolation rather than as part of a group of decisions or theories.

So if there are 20 possible explanations for a problem, and there is no strong evidence suggesting any one explanation, then I will have 5% certainty of the average explanation. Unless I am extremely good at calibration, then I can't confirm any of them, and if I consider each explanation in isolation from the other explanations, then all of them are wrong.

It doesn't matter whether we're talking about hypotheses or decision-making.

Comment author: gjm 22 December 2015 07:20:38PM 1 point [-]

Bayesian confirmation theory as a philosophy was originally conceived as a model using only two possibilities

I'm not sure whether this is true, but it's irrelevant. Bayesian confirmation theory works just fine with any number of hypotheses.

then I can't confirm any of them

If by "confirm" you mean "assign high probability to, without further evidence", yes. That seems to me to be exactly what you'd want. What is the problem you see here?

Comment author: Lumifer 22 December 2015 05:09:22PM 1 point [-]

If it had been originally conceived using more than two possibilities, it wouldn't have made any sense to use the word confirmation.

You sound confused. The "confirmation" stems from

In Bayesian Confirmation Theory, it is said that evidence confirms (or would confirm) hypothesis H (to at least some degree) just in case the prior probability of H conditional on E is greater than the prior unconditional probability of H

(source)

Comment author: FrameBenignly 22 December 2015 05:27:17PM 0 points [-]

So what if p(H) = 1, p(H|A) = .4, p(H|B) = .3, and p(H|C) = .3? The evidence would suggest all are wrong. But I have also determined that A, B, and C are the only possible explanations for H. Clearly there is something wrong with my measurement, but I have no method of correcting for this problem.

Comment author: Lumifer 22 December 2015 05:38:15PM *  2 points [-]

H is Hypothesis. You have three: HA, HB, and HC. Let's say your prior is that they are equally probable, so the unconditional P(HA) = P(HB) = P(HC) = 0.33

Let's also say you saw some evidence E and your posteriors are P(HA|E) = 0.4, P(HB|E) = 0.3, P(HC|E) = 0.3. This means that evidence E confirms HA because P(HA|E) > P(HA). This does not mean that you are required to believe that HA is true or bet your life's savings on it.

Comment author: FrameBenignly 22 December 2015 05:57:35PM 0 points [-]

That's a really good explanation of part of the problem I was getting at. But that requires considering the three hypotheses as a group rather than in isolation from all other hypotheses to calculate 0.33.

Comment author: Lumifer 22 December 2015 06:06:17PM *  1 point [-]

But that requires considering the three hypotheses as a group rather than in isolation from all other hypotheses to calculate 0.33

No, it does not.

Let's say you have a hypothesis HZ. You have a prior for it, say P(HZ) = 0.2 which means that you think that there is a 20% probability that HZ is true and 80% probability that something else is true. Then you see evidence E and it so happens that the posterior for HZ becomes 0.25, so P(HZ|E) = 0.25. This means that evidence E confirmed hypothesis HZ and that statement requires nothing from whatever other hypotheses HA,B,C,D,E,etc. might there be.

Comment author: Vaniver 22 December 2015 06:13:12PM 0 points [-]

But that requires considering the three hypotheses as a group rather than in isolation from all other hypotheses to calculate 0.33.

Not really. A hypothesis's prior probability comes from the total of all of your knowledge; in order to determine that P(HA)=0.33 Lumifer needed the additional facts that there were three possibilities that were all equally likely.

It works just as well if I say that my prior is P(HA)=0.5, without any exhaustive enumeration of the other possibilities. Then evidence E confirms HA if P(HA|E)>P(HA).

(One should be suspicious that my prior probability assessment is a good one if I haven't accounted for all the probability mass, but the mechanisms still work.)

Comment author: gjm 22 December 2015 07:24:09PM 1 point [-]

If you start with inconsistent assumptions, you get inconsistent conclusions. If you believe P(H)=1, P(A&B&C)=1, and P(H|A) etc. are all <1, then you have already made a mistake. Why are you blaming this on Bayesian confirmation theory?

Comment author: LawrenceC 22 December 2015 05:31:41PM 0 points [-]

Wait, how would you get P(H) = 1?

Comment author: FrameBenignly 22 December 2015 05:36:30PM 0 points [-]

Fine. p(H) = 0.5, p(H|A) = 0.2, p(H|B) = 0.15, p(H|C) = 0.15 It's not really relevant to the problem.

Comment author: Vaniver 22 December 2015 06:28:44PM *  0 points [-]

It's not really relevant to the problem.

The relevance is that it's a really weird way to set up a problem. If P(H)=1 and P(H|A)=0.4 then it is necessarily the case that P(A)=0. If that's not immediately obvious to you, you may want to come back to this topic after sleeping on it.

Comment author: IlyaShpitser 22 December 2015 05:38:06PM 0 points [-]

\sum_i p(H|i) need not add up to p(H) (or indeed to 1).

Comment author: passive_fist 22 December 2015 09:26:04AM 4 points [-]

A major weakness of the standard Bayesian inference method is that it assumes a problem only has two possible solutions.

This is not true at all.

Comment author: FrameBenignly 22 December 2015 03:30:10PM *  -1 points [-]

A large chunk of academics would say that it is. For example, from the paper I was referencing in my post:

At some point in history, a statistician may well write down a model which he or she believes contains all the systematic influences among properly defined variables for the system of interest, with correct functional forms and distributions of noise terms. This could happen, but we have never seen it, and in social science we have never seen anything that comes close. If nothing else, our own experience suggests that however many different specifications we thought of, there are always others which did not occur to us, but cannot be immediately dismissed a priori, if only because they can be seen as alternative approximations to the ones we made. Yet the Bayesian agent is required to start with a prior distribution whose support covers all alternatives that could be considered.

Comment author: gjm 22 December 2015 03:47:46PM 2 points [-]

That doesn't at all say Bayesian reasoning assumes only two possibilities. It says Bayesian reasoning assumes you know what all the possibilities are.

Comment author: FrameBenignly 22 December 2015 03:56:03PM *  0 points [-]

True, but how often do you see an explanation of Bayesian reasoning in philosophy that uses more than two possibilities?

Comment author: MrMind 22 December 2015 08:25:08AM *  3 points [-]

A major weakness of the standard Bayesian inference method is that it assumes a problem only has two possible solutions.

This is a weird sentence to me. I learned about Bayesian inference through Jaynes' book and surely it doesn't portray that inference as having only two possible solutions.
The other book I know about, Sivia's, doesn't do this either.

Comment author: FrameBenignly 22 December 2015 03:30:57PM 0 points [-]

You're referring to how it is described in statistics textbooks. I'm talking about confirmation theory as a philosophy.