You're looking at Less Wrong's discussion board. This includes all posts, including those that haven't been promoted to the front page yet. For more information, see About Less Wrong.

OrphanWilde comments on What's wrong with this picture? - Less Wrong Discussion

15 Post author: CronoDAS 28 January 2016 01:30PM

You are viewing a comment permalink. View the original post to see all comments and the full post content.

Comments (80)

You are viewing a single comment's thread. Show more comments above.

Comment author: OrphanWilde 01 February 2016 07:15:26PM 1 point [-]

If you don't even know what you mean by surprise (because that's what we're ostensibly trying to figure out, right?), then how can you use the math to deduce that some quantitative measure of surprise is equal in all cases?

Because the number of bits of information is the same in all cases. Any given random sequence provides evidence of countless extremely low probability world models - we just don't consider the vast majority of those world-models because they aren't elevated to our attention.

If in the future we want to keep using a coin flip example, I suggest forgetting things that are so mind-like as 'Alice is clairvoyant', and maybe just talk about biased and unbiased coins. It seems like an unnecessary complication.

It's both necessary and relevant. Indeed, I crafted my example to make your brain come up with that answer. Your conscious mind, once aware of it, probably immediately threw it into the "Silly explanation" column, and I'd hazard a guess that if asked, you'd say you wrote it down as a joke.

Because it clearly isn't an example of a world-model being allocated evidence. Your explanation is post-hoc - that is, you're rationalizing. Your description would be an elegant mathematical explanation - I just don't think it's correct, as pertains to what your mind is actually doing, and why you find some situations more surprising than others.

Comment author: Gram_Stone 02 February 2016 12:03:23AM 0 points [-]

Because the number of bits of information is the same in all cases.

I don't know why you're using self-information/surprisal interchangeably with surprise. It's confusing.

Any given random sequence provides evidence of countless extremely low probability world models - we just don't consider the vast majority of those world-models because they aren't elevated to our attention.

Like in the sense that there are hypotheses that something omniscient would consider more likely conditional on Alice doing something surprising, that humans just don't think of because they're humans? I don't expect problems coming up with a satisfactory description of 'the space of all world-models' to be something we have to fix before we can say anything important about surprise.

Because it clearly isn't an example of a world-model being allocated evidence. Your explanation is post-hoc - that is, you're rationalizing. Your description would be an elegant mathematical explanation - I just don't think it's correct, as pertains to what your mind is actually doing, and why you find some situations more surprising than others.

Maybe there's more to be said about the entire class of things that humans have ever labeled as surprising, but this does capture something of what humans mean by surprise, and we can say with particular certainty that it captures what happens in a human mind when a visual stimulus is describable as 'surprising.' The framework I described has, to my knowledge, been shown to correspond quite closely to our neuroscientific understanding of visual surprise and has been applied in machine learning algorithms that diagnose patients based on diagnostic images. There are algorithms that register seeing a tumor on a CT scan as 'surprising' in a way that is quite likely to be very similar to the way that a human would see that tumor and feel surprised. (I don't mean that it's similar in a phenomenological sense. I'm not suggesting that these algorithms have subjective experiences.) I expect this notion of surprise to be generalizable.

Comment author: OrphanWilde 03 February 2016 02:55:55PM 1 point [-]

I expect this notion of surprise to be generalizable.

Which is what I'm trying to get at. There's -something- there, more than "amount of updates to world-models". I'd guess what we call surprise has a complex relationship with the amount of updates applied to world-models, such that a large update to a single world-model is more surprising than an equal "amount" of update applied across one thousand.