PhilGoetz comments on Case study: abuse of frequentist statistics - Less Wrong
You are viewing a comment permalink. View the original post to see all comments and the full post content.
You are viewing a comment permalink. View the original post to see all comments and the full post content.
Comments (96)
Thanks for the pointer to the original paper.
Check out the title: abuse of frequentist statistics. Yes, at the end, I argue from a Bayesian perspective, but you don't have to be a Bayesian to see the structural problems with frequentist statistics as currently taught to and practiced by working scientists.
Me too. But not all papers with shoddy statistics are sent to statisticians for review. Experimental biologists in particular have a reputation for math-phobia. (Does the fact that when I saw the sample size the word "underpowered" instantly jumped into my head count as evidence that I am competent?)
Well, I don't see the structural problems. (I don't even know what a structural problem is.)
Somebody, please write a top-level post addressing this. Stop saying "Frequentists are bad" and leaving it at that. This is a great story; but it's not valid argumentation to try to convert it into an anti-frequentist tract.
I'd love to see a top-level post where someone suggests the best and/or most realistic way for scientists to do their statistics. I'm actually rather ignorant with regards to probability theory. I got a D in second semester frequentist statistics (hard teacher + I didn't go to class or try very hard on the homework) which is indicative of how little I learned in that class. I did better in my applied statistics classes.
When is it good for scientists to do null hypothesis testing?
What specifically is the "this" you want addressed? I'm not sure what its referent is.