Wind comments on A rational unfalsifyable believe - Less Wrong

1 Post author: Arielgenesis 25 July 2016 02:15AM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (46)

You are viewing a single comment's thread. Show more comments above.

Comment author: Arielgenesis 25 July 2016 05:29:30PM 0 points [-]

The idea of the story is that there are no evidence. Because I think, in real life, sometimes, there are important and relevant things with no evidence. In this case, Adam's innocence is important and relevant to Eve (for emotional and social reasons I presume), but there is no, and there will never be, evidence. Given that, saying: "If there is evidence, then the belief could be falsified." is a kind of cheating because producing new evidence is not possible anymore.

Comment author: Pimgd 26 July 2016 09:36:07AM *  1 point [-]

because producing new evidence is not possible anymore.

Okay...

So, say it turns out that, well, Eve is irrational. Somehow.

Now what? Do we go "neener-neener" at her? What's the point? What's the use that you could get out of labeling this behavior irrational?

Suppose Adam dies and is cryo-frozen. During Eve's life, there will be no resuscitation of Adam. Sometime afterward, however, Omega will arrive, deem the problem interesting and simulate Adam via really really really advanced technology.

Turns out he didn't do it.

Is she now rational because, well, turns out she was right after all? Well, no, because getting the right answer for the wrong reasons is not the rational way to go about things (in general, it might help in specific cases if you need to get the answer right but don't care how).

....

Actually, let me just skip over a few paragraphs I was going to write and skip to the end.

You cannot have 100% confidence interval. Because then your belief is set in stone and it cannot change. You can have a googleplex nines if you want, but not 100% confidence.

Fallacy of argument from probability (if it can happen then it must happen) aside; How is it rational to discard a belief you are holding on shaky evidence if you think with near absolute certainty that no more evidence will arrive, ever? What will you do when there is more evidence? (Hint: Meeting Adam's mother at the funeral and hearing childhood stories about what a nice kid he was is more evidence for his character, albeit very weak evidence - and so are studies that show that certain demographics of the timeperiod that Adam lived in had certain characteristics) You gotta update! (I don't think that fallacy I mentioned applies; if it does, we can fix it with big numbers; if you are to hold this belief everywhere, then... the probabilities go up as it turns from "in this situation" to "in at least one of all these situations")

So to toss a belief aside because you think there will be no more evidence is the wrong action to me. You can park a belief. That is to take no action. Maintain status quo. No change in input is no change in output. But you do NOT clear the belief.

Let me put up a strawman - I'll leave it up to others to see if there's something harder underneath - if you hold this action - "I think there will be no more evidence, and I am not very confident either way, so I will discard the output" to be the rightful action to take, how do you prevent yourself from getting boiled like a frog in a pan (yes, that's a false story - still, I intend the metaphorical meaning: how do you stop yourself from discarding every bit of evidence that comes your way, because you "know" there to be no more evidence?)

In my opinion, to do as you say weakens or even destroys the gradual "update" mechanism. This leads to less effective beliefs, and thus is irrational.


Were we to now look at the 3 questions, I'd answer..

Again, Eve is irrational because she says it cannot be falsified. If we let Eve say "I still think he didn't do it because of his character, and I will keep believing this until I see evidence to the contrary - and if such evidence doesn't exist, I will keep believing this forever" - then yes, Eve is rational.

The second question, yes via this specific example. Here it can, thus it can.

Yes, it can be extended to belief in God. Provided we restrict "God" to a REALLY TINY thing. As in, gee, a couple thousand years ago, something truly fantastic happened - it was God! I saw it with my own eyes! You can keep believing there was, at that point in time, an entity causing this fantastic thing. Until you get other evidence, which may never happen. What you CANNOT do is say, "hey, maybe this 'God' that caused this one fantastic thing is also responsible for creating the universe and making my neighbor win the lottery and my aunt get cancer and ..." That's unloading a huge complexity on an earlier belief without paying appropriate penalties.

You don't only need evidence that the fantastical events were caused, you also need evidence they were caused by the same thing if you wish to attribute them to that same thing.

Comment author: Wind 28 July 2016 11:36:56AM 0 points [-]

You don't only need evidence that the fantastical events were caused, you also need evidence they were caused by the same thing if you wish to attribute them to that same thing.

Assume I observe X, Y, Z and form three hypotheses

  • A: All of X, Y, Z had causes
  • B: All of X, Y, Z had different causes
  • C: All of X, Y, Z had the same cause

A obviously has highest probability since it includes B and C as special cases. However, which one of B and C do you think should get complexity penalty over the others?

In you story:

Yes, it can be extended to belief in God. Provided we restrict "God" to a REALLY TINY thing. As in, gee, a couple thousand years ago, something truly fantastic happened - it was God! I saw it with my own eyes! You can keep believing there was, at that point in time, an entity causing this fantastic thing. Until you get other evidence, which may never happen. What you CANNOT do is say, "hey, maybe this 'God' that caused this one fantastic thing is also responsible for creating the universe and making my neighbor win the lottery and my aunt get cancer and ..." That's unloading a huge complexity on an earlier belief without paying appropriate penalties.

The relevant comparison is: Given that God did X, what is the probability that God also did Y and Z, verses God did not do those things.

P(God did Y, Z | God did X) = P(God did X,Y, Z) / P(God did X)

v.s.

P(God did not do Y, Z | God did X) = P(God did X, and something other than God did Y, Z) / P(God did X)

I am uncertain about how to correctly apply complexity penalty, but I do believe that the multi explanation model "God did X, and something other than God did Y, Z" should get complexity penalty over the sing explanation model "God did X, Y, Z".

The belief "God caused some tiny thing, a couple thousand years ago", should correlated with the belief "God did this big thing right now". This is why I firmly believe that God did not cause some tiny thing, a couple of thousand years ago.

Comment author: Pimgd 28 July 2016 03:55:57PM 0 points [-]

Phrased like this, I see what you're getting at; but in my mind, I was describing extraordinary, but different events. Say, miracle cures and miracle plagues or whatever. A whole bunch of locusts and your aunt being cured of cancer most likely have different causes. In that case, you first have to postulate an entity which can summon a bunch of locusts. The actual summoning need not be magical or spontaneous in nature, only their appearance. So keeping a bunch of locusts hidden away whilst feeding them (somehow), before releasing them like a plague, would do.

This SAME entity then also needs the ability to cure cancer. To me, adding abilities like this incurs complexity penalties on a pretty big scale. Especially when you start adding other stuff and start scaling this influence over time (same entity responsible for actions many thousands of years ago and events now)

Comment author: Lumifer 28 July 2016 05:33:24PM *  1 point [-]

This SAME entity then also needs the ability to cure cancer. To me, adding abilities like this incurs complexity penalties on a pretty big scale.

This says that if you are, say, an Inca ruler and you hear about Spanish conquistadors, the fact that they can ride weird beasts AND shoot fire out of metal sticks AND do a lot of other supernatural-looking stuff implies that you should disbelieve their existence -- probably not a good idea.

In general terms, the complexity penalties you're are talking about are justified only if these different abilities are unrelated. But if, instead, all of them have a common cause (e.g. massive technological superiority), the penalties no longer apply.

Comment author: Pimgd 29 July 2016 07:20:10AM 0 points [-]

I see.