Matt_Simpson comments on Open Thread: July 2010 - Less Wrong
You are viewing a comment permalink. View the original post to see all comments and the full post content.
You are viewing a comment permalink. View the original post to see all comments and the full post content.
Comments (653)
They most certainly are. But it's semantics.
Frankly, I'm not informed enough about priors commit to maxent, Kolmogorov complexity, or anything else.
yes
aaahhh.... I changed the language of that sentence at least three times before settling on what you saw. Here's what I probably should have posted (and what I was going to post until the last minute):
That is probably intuitively easier to grasp, but I think a bit inconsistent with my language in the rest of the post. The language is somewhat difficult here because our uncertainty is simultaneously a map and a territory.
For the record, I thought this sentence was perfectly clear. But I am a statistics grad student, so don't consider me representative.
Are you asserting that this a catch for my position? Or the "never look back" approach to priors? What you are saying seems to support my argument.
OK. I agree with that insofar as agents having the same prior entails them having the same model.
Ah, I think I get you; a PB (perfect Bayesian) doesn't see a need to test their model because whatever specific proposition they're investigating implies a particular correct model.
Yeah, I figured you wouldn't have trouble with it since you talked about taking classes in this stuff - that footnote was intended for any lurkers who might be reading this. (I expected quite a few lurkers to be reading this given how often the Gelman and Shalizi paper's been linked here.)
It's a catch for the latter, the PB. In reality most scientists typically don't have a wholly unambiguous proposition worked out that they're testing - or the proposition they are testing is actually not a good representation of the real situation.