Cyan comments on Frequentist Statistics are Frequently Subjective - Less Wrong
You are viewing a comment permalink. View the original post to see all comments and the full post content.
You are viewing a comment permalink. View the original post to see all comments and the full post content.
Comments (81)
Damn. That is some wrong-headed thinking right there. Never mind the likelihood principle -- have these people never heard of cross-validation?
I don't follow; could you explain? It seems like a decent idea to me.
The fear seems to be that people might propose theories with enough degrees of freedom that they can fine-tune it to fit the complete data very closely. But, as long as the fitting process is repeatable, i.e., no numerology, cross-validation can be applied to discover which theories are genuinely predictive and which are over-fitting.
OK; so I'm looking at http://en.wikipedia.org/wiki/Cross-validation_%28statistics%29 and the key nugget seems to be:
But how can we come up with a hypothetical validation set for the Universe?
Some of the lines seem to suggest exactly what those researchers seem to propose - not allowing all of the observations to go into the theory, but like the NetFlix Prize, holding back some data as a test:
Releasing the data in dribs and drabs doesn't address this either.
There's a difference between, on the one hand, having the data freely available and being intelligent enough to use cross-validation, and on the other, having someone paternalistically hold the data back from you.
It does force researchers into an ad hoc cross-validation scheme, doesn't it?
If you start from the premise that researchers may fall into the overfitting trap, then you're already treating them adversarily. And if just one researcher overfitting a theory and so becoming irrefutable will screw everything up, then the paranoid approach to data release prevents that total cockup (at the cost of some interim inefficiencies, by hindering the responsible, good, researchers).
It doesn't prevent that entirely reliably, either. How much time are you going to give the researchers to come up with hypotheses before you release the full set? And what do you do if someone comes up with a new hypothesis after the full release, so mindblowingly elegant and simple that it blows all of the previously published ones out of the water?
If you think that some later hypotheses based on the full set might still be accepted, then you're still vulnerable to falling into the overfitting trap after the full release. If you don't, then you'll be locked forever into the theories scientists came up with during the partial-release window, and no later advances in the scientific method, rationality, computing, math or even the intelligence of researchers will allow you to improve upon them.
This approach might get you some extra empirical evidence, but it will be empirical evidence about theories put together under quite limited conditions, compared to what will be available to later civilization.
I'd rather wait for researchers to screw up and then hammer them.