Yeah, I basically see this episode as anti-science propaganda.
The "friendship lesson" basically says "make-belief is a good thing and should be respected".
Either that, or accepting the "supernatural" as such without further inquiry. Because it's by definition beyond the realm of science, duh.
(Whether it's intentional anti-science-propaganda is another question)
You could use the "zombie argument" to "prove" that any kind of machine is more than the sum of its parts.
For example, imagine a "zombie car" which is the same on an atom-by-atom basis as a normal car, except it doesn't drive.
In this context, the absurdity of the zombie argument should be more obvious.
EDIT: OK, it isn't quite the same kind of argument, since the car wouldn't behave exactly the same, but it's pretty similar.
EDIT2: Another example to illustrate the absurdity of the zombie argument:
You could imagine an alternative world that's exactly t...
"Regarding the first question: evolution hasn’t made great pleasure as accessible to us as it has made pain. Fitness advantages from things like a good meal accumulate slowly but a single injury can drop one’s fitness to zero, so the pain of an injury is felt stronger than the joy of pizza. But even pizza, though quite an achievement, is far from the greatest pleasure imaginable.
Humankind has only recently begun exploring the landscape of bliss, compared to our long evolutionary history of pain. If you can’t imagine a pleasure great enough to make the trad...
If I understand correctly, you may also reach your position without using a of non-causal decision theory if you mix utilitarianism with the deontological constraint of being honest (or at least meta-honest [see https://www.lesswrong.com/posts/xdwbX9pFEr7Pomaxv/meta-honesty-firming-up-honesty-around-its-edge-cases]) about the moral decisions you would make.
If people would ask you whether you would kill/did kill a patient, and you couldn't confidently say "No" (because of the deontological constraint of (meta-)honesty), that would be pretty bad, so you must...
slighly modified version:
Instead of chosing at once whether you want to take one box or both boxes, you first take box 1 (and see whether it includes 0$ or 1.000.000$), and then, you decide whether you want to also take box 2.
Assume that you only care about the money, you don't care about doing the opposite of what Omega predicted.
slightly related:
Suppose Omega forces you to chose a number 0<p<=1 and then, with probability p, you get tortured for 1/(p²) seconds.
Assume for any T, being tortured for 2T seconds is exactly twice as bad as being tortured for T seconds.
Also assume that your memory gets erased afterwards (this is to make sure there won't be additional suffering from something like PTSD)
The expected value of seconds being tortured is p * 1/(p²)=1/p, so, in terms of expected value, you should chose p=1 and be tortured for 1 second. The smaller the p you chose, the higher the expected value.
Would you actually chose p=1 to maximize the expected value, or would you rather chose a very low p (like 1/3^^^^3)?
I once thought I could prove that the set of all natural numbers is as large as its power set. However, I was smart enough to acknowledge my limitations (What‘s more likely: That I made a mistake in my thinking I haven‘t yet noticed, or that a theorem pretty much any professional mathematician accepts as true is actually false?), so I activly searched for errors in my thinking. Eventually, I noticed that my methods only works for finite sub sets (The set of all natural numbers is, indeed, as large as the set of all FINITE subsets), but not for infinite subsets.
Eliziers method also works for all finite subsets, but not for infinite subsets
My answers:
1.No, because their belief doesn't make any sense. It even has logical contradictions, which makes it "super impossible", meaning there's no possible world where it could be true (the omnipotence paradox proves that omnipotence is logically inconsistent; a god which is nearly omnipotent, nearly omniscient and nearly omnibenevolent wouldn't allow suffering, which, undoubtably, exists; "God wants to allow free will" isn't a valid defence, since there's a lot of suffering that isn't caused by other ...
There would actually be several changes:
I would stop being vegan.
I would stop donating money (note: I currently donate quite a lot of money for projects of "Effective altruism").
I would stop caring about Fairtrade.
I would stop feeling guilty about anything I did, and stop making any moral considerations about my future behaviour.
If others are overly friendly, I would fully abuse this for my advantage.
I might insult or punch strangers "for fun" if I'm pretty sure I will never see them again (and they don't seem like the ...
More acuratly, "absence of evidence you would expect to see if the statement is true" is evidence of absence.
If there's no evidence you'd expect if the statement is true, absence of evidence is not evidence of absence.
For example, if I tell you I've eaten cornflakes for breakfast, no matter whether or not the statement is true, you won't have any evidence in either direction (except for the statement itself) unless you're willing to investigate the matter (like, asking my roommates). In this case, absence of evidence is n...
I see another way to show that 1/5 is the correct solution:
P(2 Aces | Ace of Spades revealed)= P(2 Aces AND Ace of Spades revealed)/P(Ace of Spades revealed)
(note: for further calculations, I'm assuming that there are 5 possible hands and the probability for each hand is 1/5, since it already has been revealed that there is at least one Ace. The end result would be the same if you would also set aside a random card in case you have no Ace,but the probabilities in the steps before the end results would have to change accordingly)
P(2 Aces AND Ace of Spades reveled)=P(2 Aces)*1/2 = 1/5 * 1/2 =1/10
P(Ace of Spades revealed)= 2/5 * 1 + 1/5 * 1/2 = 5/10
(1/10)/(5/10)=1/5
Assigning Bayes-probabilities <1 to mathematical statements (that have been definitly proven) seems absurd and logically contradictory, because you need mathematics to even asign probabilities.
If you assign any Bayes probability to the statement that Bayes probabilities even work, you already assume that they do work.
And, arguably, 2+2=4 is much simpler than the concept of Bayes-probability (To be fair, the same might not be true for my most complex statement that Pi is irrational)
I agree that you can never be „infinitly certain“ about the way the physical world is (because there‘s always a very tiny possibility that things might suddenly change, or everything is just a simulation, or a dream, or […] ), but you should assign probability 1 to mathematical statements for which there isn‘t just evidence, but actual, solid proof.
Suppose you have the choice beetween the following options: A You get a lottery with a 1-Epsilon chance of winning. B You win if 2+2=4 and 53 is a prime number and Pi is an irrational number.
Is there any Ep
...Suppose I have two cards, A and B, that I shuffle and then blindly place in two spaceships, pointed at opposite ends of the galaxy. If they go quickly enough, it can be the case that they get far enough apart that they will never be able to meet again. But if you're in one of the spaceships, and turn the card over to learn that it's card A, then you learn something about the world on the other side of the light cone boundary.
Jeez, „Collapse of Western Civilisation“, that‘s some serious clickbait.