When degrees of freedom in explanatory flexibility go up people wind up cold reading themselves, otherwise known as synchronicity.
I wrote the following comment over there which seemed to be caught in the spam filter or something:
Both Sue and the third party are rational, and them knowing all objective facts about everyone's experiences does not eliminate the disagreement.
The reason is because the evidence is anthropic in nature - it is more likely under certain hypotheses that affect the probability of "you" existing or "you" having certain experiences, above and beyond objective facts. Such evidence is agent-centered.
For example, Sue's evidence raises the probability of the hypothesis "God exists and cares about me in particular" for her, but not for the third party. Of course, the third party's probability of "God cares about Sue in particular" goes up. But that has a lower prior probability when it's about someone else, because that hypothesis also predicts that "I" will "be Sue" more than the baseline expectation of 1 in 7 billion or so.
In general, the class of hypotheses that Sue's evidence favors also tends to make Sue's existence and sentience more likely. Since Sue knows that she exists and is sentient but does not know that about anyone else, she starts with a higher prior probability in that class of hypotheses, and therefore the same update as third parties will result in a higher posterior probability for that class.
This also means that Sue's family and barista rationally conclude somewhat weaker versions of the class of hypotheses - their priors should be higher than a random third party, but lower than Sue's priors.
This is analogous to the argument with lottery winners here: https://www.lesswrong.com/posts/kKAmxmQq9umJiMFSp/real-life-anthropic-weirdness
But what is your watching friend supposed to think? Though his predicament is perfectly predictable to you - that is, you expected before starting the experiment to see his confusion - from his perspective it is just a pure 100% unexplained miracle. What you have reason to believe and what he has reason to believe would now seem separated by an uncrossable gap, which no amount of explanation can bridge. This is the main plausible exception I know to Aumann's Agreement Theorem.
Pity those poor folk who actually win the lottery! If the hypothesis "this world is a holodeck" is normatively assigned a calibrated confidence well above 10-8, the lottery winner now has incommunicable good reason to believe they are in a holodeck. (I.e. to believe that the universe is such that most conscious observers observe ridiculously improbable positive events.)
Your example with Sue is the same, just at a smaller scale with less evidence and therefore less strong conclusions.
Re your ethics example, you're assuming that knowledge of others' intuitions counts as moral evidence. Even if that were the case, knowledge of a single person's intuition is plausibly not enough to shift from uncertain to confident in one position, or vice versa.
Hmm... I suppose the psychic abilities hypothesis might be indirectly tied to anthropic considerations, but it feel to me as reading too much into the particular example chosen. Or maybe not, maybe almost any hypothesis has anthropic effects when it comes down to it, but I think the idea was to try to isolate one particular effect.
I think that the only rational reason to treat your own experience as more significant, given the constraints of the problem, is the anthropic nature of the evidence. And that explains nicely why others shouldn't update as much.
If you think there's an example that isolates non-anthropic effects that behave similarly, perhaps. I'll reserve judgement until I see such an example - for now, I don't know of any.
Epistemic Status: Just some thoughts off the top of my head
Fake Nous recently featured an article on agent-centered evidence:
I actually think that there is a form of agent-centered evidence called intuition which isn't easily communicatable to other agents and can often prove fairly useful. However, that's not the issue I want to discuss today. Instead, I want to talk about reference classes. But before I cover that, let's talk about shots. Suppose we have a raffle with one prize and a thousand tickets. If you have one ticket, you have one shot, while if you have ten tickets, you have ten shots. I haven't precisely defined it, but I think this example should be clear enough. Once you know the number of shots, you can turn it into a probability.
Let's suppose that you have four members in your family and that you also have four work colleagues. If one of your family members has a precognition-like experience, you might say that it is remarkable as you only had four shots, but let's suppose that in the counterfactual where one of your colleagues had an experience you would have also counted it as four shots. This seems like a mistake; only one can be counted as four shots and if the other occurs, then it has to be counted as eight shots with the group being classed as family AND colleagues.
If you have a bunch of different groups, say three family members, another five work colleagues and eight cousins, then you can order them arbitrarily. It doesn't matter if you do (3,5,8) or (5,3,8) or (8,3,5); any of them is fine. If you order them (8,3,5), then the number of shots for a member of a group is 8 for the first, 11 for the second and 16 for the third.
This might seem strange. We are calculating a different probability of psychic powers existing when an event happens to a member of your family vs. one of your colleagues, even though there doesn't seem to be any fundamental reason written into the universe itself why one group should give you more evidence.
Then again, the probability you assign to something existing is really more about your subjective state, such as the information available to you and how it was generated, rather than the objective state of the universe itself. We can think about choosing how to order your groups the same way that we think about committing to a (frequentialist) experiment design in advance. It's well understood that if you test more hypotheses, you increase your chance of a spurious result. For example, if you test for effects in male adults, female adults, male children and female children, you've taken four shots as opposed to having only tested for an overall effect. This is typically adjusted for by using a significance threshold that isn't constant, but instead depends on the number of hypotheses that you are testing.
We can take this analogy further. Suppose in the example above, we choose the ordering, (5, 3, 8). Then this defines three experiments - work collegues only with five shots, work + family with eight shots and all groups with sixteen shots. If we observe one of our family members having such an experience, we can treat it as us having pre-committed to an experiment covering family and work colleagues with eight shots. This is far better than the naive tendency we might have to define the group as just family with three shots.
However, it still isn't a completely accurate way to handle probability, as if we want an accurate an estimate of psychic ability as possible, then we should take into account all the evidence available. So if we also know about whether our cousins have had such experiences, then we really should take that into account when calculating the probability. Of course, trying to figure out this implicit sample might greatly complicate the calculation, which is why this group based approximation is much more appealing instead.
That said, this is quite an unusual approximation, as it can result in completely different probabilities than if we had the whole data. For example, observing a positive out of five shots striking instead of a positive out of sixteen makes a huge difference in the actual probabilities. Nonetheless, if you had precommitted to making a decision based on the first five, then the increase in probability when you saw a positive result would be perfectly matched by the decrease when you a negative result. This means that deciding in advance to only look at the first five wouldn't bias the result, even if throws away data.
Perhaps a more realistic scenario is one where you precommit to expanding the experimental group until you hit a positive result or you've expanded it to the end. This would represent the fact that someone might not worry about how amazing it is that someone in their church had a particular experience if someone in their family had such an experience. These expansionary scenarios are too complex to handle using the shots framework, but even in this scenario the maths isn't too hard.
Of course, a lot of the time we aren't deciding in advance, but are instead deciding after the fact. In this case, you're ability to use these schemes is highly, highly dependent on your ability to self-model. If you can do this well, then you can adopt these schemes after the fact, but if you do it poorly, it'll completely mess up the results.