gwern comments on The Logic of the Hypothesis Test: A Steel Man - Less Wrong
You are viewing a comment permalink. View the original post to see all comments and the full post content.
You are viewing a comment permalink. View the original post to see all comments and the full post content.
Comments (37)
One interesting thing here is that you start with a null vs other hypothesis, but that's because you're doing a two-sample z/t-test. But what's going on when you do a one sample z/t-test and get out a confidence interval?
The one sided hypothesis test is still null vs. other because it uses the full parameter space, i.e. it's H0: mu <= c vs. Ha: mu > c. We present it to undergrads as H0: mu = c vs. Ha: mu > c in order to simplify (I think that's the reason anyway) but really we're testing the former. The Karlin-Rubin theorem justifies this.
I don't follow... that sounds like you're giving the definition of a one-tailed hypothesis test. What does that have to do with a constant c? Suppose I do this in R:
And get a 95% CI of (-0.3138-0.4668); if my null hypothesis (H0) is my mu or sample mean (0.07652), then you say my Ha is mu > c, or 0.07652 > c. What is this c?
So rereading your first comment, I realize you said one-sample vs. two-sample hypothesis test and not one-sided vs. two-sided (ore one-tailed vs. two-tailed). If that's what you meant, I don't follow your first comment. The t-test I gave in the post is a one-sample test - and I don't understand how the difference between the two is relevant here.
But to answer your question anyway:
c is the value you're testing as the null hypothesis. In that R-code, R assumes that c=0 so that H0: mu=c and Ha: mu=/=c. For the R code:
You perform a t test with H0: mu<=c and Ha: mu>c.
I'm interested in the calculated confidence interval, not the p-value necessarily. Noodling around some more, I think I'm starting to understand it more: the confidence interval isn't calculated with respect to the H0 of 0 which the R code defaults to, it's calculated based purely on the mean (and then an H0 of 0 is assumed to spit out some p-value)
Hm... I'm trying to fit this assumption into your framework....
A 95% confidence interval is sort of like testing H0:mu=c vs Ha:mu=\=c for all values of c at the same time. In fact if you reject the null hypothesis for a given c when c is outside your calculated confidence interval and fail to reject otherwise, you're performing the exact same t-test with the exact same rejection criteria as the usual one (that is if the p-value is less than 0.05).
The formula for the test statistic is (generally) t = (estimate - c)/(standard error of estimate) while the formula for a confidence interval is (generally) estimate +/- t^(standard error of estimate) where t^ is a quantile of the t distribution with appropriate degrees of freedom, chosen according to your desired confidence level. t^* and the threshold for rejecting the null in a hypothesis test are intimately related. If you google "confidence intervals and p values" I'm sure you'll find a more polished and detailed explanation of this than mine.